File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: New Problems in Distributed Inference for DNN Models on Robotic IoT

TitleNew Problems in Distributed Inference for DNN Models on Robotic IoT
Authors
Keywordsdistributed inference
distributed system and network
robotic IoT
Issue Date17-Jun-2024
PublisherACM
Abstract

The rapid advancements in machine learning (ML) techniques have led to significant achievements in various robotic tasks. Deploying these ML approaches on real-world robots requires fast and energy-efficient inference of their deep neural network (DNN) models. To our knowledge, distributed inference, which involves inference across multiple powerful GPU devices, has emerged as a promising optimization to improve inference performance in modern data centers. However, when deployed on real-world robots, existing parallel methods can not simultaneously meet the robots' latency and energy requirements and raise significant challenges.

This paper reveals and evaluates the problems hindering the application of these parallel methods in robotic IoT, including the failure of data parallelism, the unacceptable communication overhead of tensor parallelism, and the significant transmission bottlenecks in pipeline parallelism. By raising awareness of these new problems, we aim to stimulate research toward finding a new parallel method to achieve fast and energy-efficient distributed inference in robotic IoT.


Persistent Identifierhttp://hdl.handle.net/10722/348075

 

DC FieldValueLanguage
dc.contributor.authorSun, Zekai-
dc.contributor.authorGuan, Xiuxian-
dc.contributor.authorWang, Junming-
dc.contributor.authorLiu, Fangming-
dc.contributor.authorCui, Heming-
dc.date.accessioned2024-10-04T00:31:17Z-
dc.date.available2024-10-04T00:31:17Z-
dc.date.issued2024-06-17-
dc.identifier.urihttp://hdl.handle.net/10722/348075-
dc.description.abstract<p>The rapid advancements in machine learning (ML) techniques have led to significant achievements in various robotic tasks. Deploying these ML approaches on real-world robots requires fast and energy-efficient inference of their deep neural network (DNN) models. To our knowledge, distributed inference, which involves inference across multiple powerful GPU devices, has emerged as a promising optimization to improve inference performance in modern data centers. However, when deployed on real-world robots, existing parallel methods can not simultaneously meet the robots' latency and energy requirements and raise significant challenges.</p><p>This paper reveals and evaluates the problems hindering the application of these parallel methods in robotic IoT, including the failure of data parallelism, the unacceptable communication overhead of tensor parallelism, and the significant transmission bottlenecks in pipeline parallelism. By raising awareness of these new problems, we aim to stimulate research toward finding a new parallel method to achieve fast and energy-efficient distributed inference in robotic IoT.</p>-
dc.languageeng-
dc.publisherACM-
dc.relation.ispartofApPLIED'24: 2024 Workshop on Advanced Tools, Programming Languages, and PLatforms for Implementing and Evaluating algorithms for Distributed systems (17/06/2024-17/06/2024, Nantes)-
dc.subjectdistributed inference-
dc.subjectdistributed system and network-
dc.subjectrobotic IoT-
dc.titleNew Problems in Distributed Inference for DNN Models on Robotic IoT-
dc.typeConference_Paper-
dc.identifier.doi10.1145/3663338.3665828-
dc.identifier.scopuseid_2-s2.0-85198649132-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats