File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TNET.2023.3279512
- Scopus: eid_2-s2.0-85161604175
- WOS: WOS:001006077400001
- Find via

Supplementary
- Citations:
- Appears in Collections:
Article: Accelerating DNN Inference With Reliability Guarantee in Vehicular Edge Computing
| Title | Accelerating DNN Inference With Reliability Guarantee in Vehicular Edge Computing |
|---|---|
| Authors | |
| Keywords | Approximation algorithms Computational modeling Data models DNN inference acceleration mobility-aware offloading overlapped partitioning Peer-to-peer computing Reliability reliability guarantee Resource management Task analysis Vehicular edge computing |
| Issue Date | 1-Jun-2023 |
| Publisher | Institute of Electrical and Electronics Engineers |
| Citation | IEEE/ACM Transactions on Networking, 2023 How to Cite? |
| Abstract | This paper explores on accelerating Deep Neural Network (DNN) inference with reliability guarantee in Vehicular Edge Computing (VEC) by considering the synergistic impacts of vehicle mobility and Vehicle-to-Vehicle/Infrastructure (V2V/V2I) communications. First, we show the necessity of striking a balance between DNN inference acceleration and reliability in VEC, and give insights into the design rationale by analyzing the features of overlapped DNN partitioning and mobility-aware task offloading. Second, we formulate the Cooperative Partitioning and Offloading (CPO) problem by presenting a cooperative DNN partitioning and offloading scenario, followed by deriving an offloading reliability model and a DNN inference delay model. The CPO is proved as NP-hard. Third, we propose two approximation algorithms, i.e., Submodular Approximation Allocation Algorithm (SA(3)) and Feed Me the Rest algorithm (FMtR). In particular, SA(3) determines the edge allocation in a centralized way, which achieves 1/3-optimal approximation on maximizing the inference reliability. On this basis, FMtR partitions the DNN models and offloads the tasks to the allocated edge nodes in a distributed way, which achieves 1/2-optimal approximation on maximizing the inference reliability. Finally, we build the simulation model and give a comprehensive per-formance evaluation, which demonstrates the superiority of the proposed solutions. |
| Persistent Identifier | http://hdl.handle.net/10722/357014 |
| ISSN | 2023 Impact Factor: 3.0 2023 SCImago Journal Rankings: 2.034 |
| ISI Accession Number ID |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Liu, K | - |
| dc.contributor.author | Liu, CH | - |
| dc.contributor.author | Yan, GZ | - |
| dc.contributor.author | Lee, VCS | - |
| dc.contributor.author | Cao, JN | - |
| dc.date.accessioned | 2025-06-23T08:52:56Z | - |
| dc.date.available | 2025-06-23T08:52:56Z | - |
| dc.date.issued | 2023-06-01 | - |
| dc.identifier.citation | IEEE/ACM Transactions on Networking, 2023 | - |
| dc.identifier.issn | 1063-6692 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/357014 | - |
| dc.description.abstract | This paper explores on accelerating Deep Neural Network (DNN) inference with reliability guarantee in Vehicular Edge Computing (VEC) by considering the synergistic impacts of vehicle mobility and Vehicle-to-Vehicle/Infrastructure (V2V/V2I) communications. First, we show the necessity of striking a balance between DNN inference acceleration and reliability in VEC, and give insights into the design rationale by analyzing the features of overlapped DNN partitioning and mobility-aware task offloading. Second, we formulate the Cooperative Partitioning and Offloading (CPO) problem by presenting a cooperative DNN partitioning and offloading scenario, followed by deriving an offloading reliability model and a DNN inference delay model. The CPO is proved as NP-hard. Third, we propose two approximation algorithms, i.e., Submodular Approximation Allocation Algorithm (SA(3)) and Feed Me the Rest algorithm (FMtR). In particular, SA(3) determines the edge allocation in a centralized way, which achieves 1/3-optimal approximation on maximizing the inference reliability. On this basis, FMtR partitions the DNN models and offloads the tasks to the allocated edge nodes in a distributed way, which achieves 1/2-optimal approximation on maximizing the inference reliability. Finally, we build the simulation model and give a comprehensive per-formance evaluation, which demonstrates the superiority of the proposed solutions. | - |
| dc.language | eng | - |
| dc.publisher | Institute of Electrical and Electronics Engineers | - |
| dc.relation.ispartof | IEEE/ACM Transactions on Networking | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.subject | Approximation algorithms | - |
| dc.subject | Computational modeling | - |
| dc.subject | Data models | - |
| dc.subject | DNN inference acceleration | - |
| dc.subject | mobility-aware offloading | - |
| dc.subject | overlapped partitioning | - |
| dc.subject | Peer-to-peer computing | - |
| dc.subject | Reliability | - |
| dc.subject | reliability guarantee | - |
| dc.subject | Resource management | - |
| dc.subject | Task analysis | - |
| dc.subject | Vehicular edge computing | - |
| dc.title | Accelerating DNN Inference With Reliability Guarantee in Vehicular Edge Computing | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1109/TNET.2023.3279512 | - |
| dc.identifier.scopus | eid_2-s2.0-85161604175 | - |
| dc.identifier.eissn | 1558-2566 | - |
| dc.identifier.isi | WOS:001006077400001 | - |
| dc.publisher.place | PISCATAWAY | - |
| dc.identifier.issnl | 1063-6692 | - |
