File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1145/3663338.3665828
- Scopus: eid_2-s2.0-85198649132
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: New Problems in Distributed Inference for DNN Models on Robotic IoT
Title | New Problems in Distributed Inference for DNN Models on Robotic IoT |
---|---|
Authors | |
Keywords | distributed inference distributed system and network robotic IoT |
Issue Date | 17-Jun-2024 |
Publisher | ACM |
Abstract | The rapid advancements in machine learning (ML) techniques have led to significant achievements in various robotic tasks. Deploying these ML approaches on real-world robots requires fast and energy-efficient inference of their deep neural network (DNN) models. To our knowledge, distributed inference, which involves inference across multiple powerful GPU devices, has emerged as a promising optimization to improve inference performance in modern data centers. However, when deployed on real-world robots, existing parallel methods can not simultaneously meet the robots' latency and energy requirements and raise significant challenges. This paper reveals and evaluates the problems hindering the application of these parallel methods in robotic IoT, including the failure of data parallelism, the unacceptable communication overhead of tensor parallelism, and the significant transmission bottlenecks in pipeline parallelism. By raising awareness of these new problems, we aim to stimulate research toward finding a new parallel method to achieve fast and energy-efficient distributed inference in robotic IoT. |
Persistent Identifier | http://hdl.handle.net/10722/348075 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Sun, Zekai | - |
dc.contributor.author | Guan, Xiuxian | - |
dc.contributor.author | Wang, Junming | - |
dc.contributor.author | Liu, Fangming | - |
dc.contributor.author | Cui, Heming | - |
dc.date.accessioned | 2024-10-04T00:31:17Z | - |
dc.date.available | 2024-10-04T00:31:17Z | - |
dc.date.issued | 2024-06-17 | - |
dc.identifier.uri | http://hdl.handle.net/10722/348075 | - |
dc.description.abstract | <p>The rapid advancements in machine learning (ML) techniques have led to significant achievements in various robotic tasks. Deploying these ML approaches on real-world robots requires fast and energy-efficient inference of their deep neural network (DNN) models. To our knowledge, distributed inference, which involves inference across multiple powerful GPU devices, has emerged as a promising optimization to improve inference performance in modern data centers. However, when deployed on real-world robots, existing parallel methods can not simultaneously meet the robots' latency and energy requirements and raise significant challenges.</p><p>This paper reveals and evaluates the problems hindering the application of these parallel methods in robotic IoT, including the failure of data parallelism, the unacceptable communication overhead of tensor parallelism, and the significant transmission bottlenecks in pipeline parallelism. By raising awareness of these new problems, we aim to stimulate research toward finding a new parallel method to achieve fast and energy-efficient distributed inference in robotic IoT.</p> | - |
dc.language | eng | - |
dc.publisher | ACM | - |
dc.relation.ispartof | ApPLIED'24: 2024 Workshop on Advanced Tools, Programming Languages, and PLatforms for Implementing and Evaluating algorithms for Distributed systems (17/06/2024-17/06/2024, Nantes) | - |
dc.subject | distributed inference | - |
dc.subject | distributed system and network | - |
dc.subject | robotic IoT | - |
dc.title | New Problems in Distributed Inference for DNN Models on Robotic IoT | - |
dc.type | Conference_Paper | - |
dc.identifier.doi | 10.1145/3663338.3665828 | - |
dc.identifier.scopus | eid_2-s2.0-85198649132 | - |