File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Hierarchical Deep Neural Network Inference for Device-Edge-Cloud Systems

TitleHierarchical Deep Neural Network Inference for Device-Edge-Cloud Systems
Authors
Keywordsdistributed systems
edge computing
neural networks
Issue Date2023
Citation
ACM Web Conference 2023 - Companion of the World Wide Web Conference, WWW 2023, 2023, p. 302-305 How to Cite?
AbstractEdge computing and cloud computing have been utilized in many AI applications in various fields, such as computer vision, NLP, autonomous driving, and smart cities. To benefit from the advantages of both paradigms, we introduce HiDEC, a hierarchical deep neural network (DNN) inference framework with three novel features. First, HiDEC enables the training of a resource-adaptive DNN through the injection of multiple early exits. Second, HiDEC provides a latency-aware inference scheduler, which determines which input samples should exit locally on an edge device based on the exit scores, enabling inference on edge devices with insufficient resources to run the full model. Third, we introduce a dual thresholding approach allowing both easy and difficult samples to exit early. Our experiments on image and text classification benchmarks show that HiDEC significantly outperforms existing solutions.
Persistent Identifierhttp://hdl.handle.net/10722/343423

 

DC FieldValueLanguage
dc.contributor.authorIlhan, Fatih-
dc.contributor.authorTekin, Selim Furkan-
dc.contributor.authorHu, Sihao-
dc.contributor.authorHuang, Tiansheng-
dc.contributor.authorChow, Ka Ho-
dc.contributor.authorLiu, Ling-
dc.date.accessioned2024-05-10T09:08:01Z-
dc.date.available2024-05-10T09:08:01Z-
dc.date.issued2023-
dc.identifier.citationACM Web Conference 2023 - Companion of the World Wide Web Conference, WWW 2023, 2023, p. 302-305-
dc.identifier.urihttp://hdl.handle.net/10722/343423-
dc.description.abstractEdge computing and cloud computing have been utilized in many AI applications in various fields, such as computer vision, NLP, autonomous driving, and smart cities. To benefit from the advantages of both paradigms, we introduce HiDEC, a hierarchical deep neural network (DNN) inference framework with three novel features. First, HiDEC enables the training of a resource-adaptive DNN through the injection of multiple early exits. Second, HiDEC provides a latency-aware inference scheduler, which determines which input samples should exit locally on an edge device based on the exit scores, enabling inference on edge devices with insufficient resources to run the full model. Third, we introduce a dual thresholding approach allowing both easy and difficult samples to exit early. Our experiments on image and text classification benchmarks show that HiDEC significantly outperforms existing solutions.-
dc.languageeng-
dc.relation.ispartofACM Web Conference 2023 - Companion of the World Wide Web Conference, WWW 2023-
dc.subjectdistributed systems-
dc.subjectedge computing-
dc.subjectneural networks-
dc.titleHierarchical Deep Neural Network Inference for Device-Edge-Cloud Systems-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1145/3543873.3587370-
dc.identifier.scopuseid_2-s2.0-85159571402-
dc.identifier.spage302-
dc.identifier.epage305-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats