File Download
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/FCCM57271.2023.00018
- Scopus: eid_2-s2.0-85165963734
- WOS: WOS:001032713500009
Supplementary
- Citations:
- Appears in Collections:
Conference Paper: Model-Platform Optimized Deep Neural Network Accelerator Generation through Mixed-Integer Geometric Programming
| Title | Model-Platform Optimized Deep Neural Network Accelerator Generation through Mixed-Integer Geometric Programming |
|---|---|
| Authors | |
| Keywords | FPGA hardware generation machine learning |
| Issue Date | 1-Jul-2023 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Abstract | Although there are distinct power-performance advantages in customizing an accelerator for a specific combination of FPGA platform and neural network model, developing such highly customized accelerators is a challenging task due to the massive design space spans from the range of network models to be accelerated, the target platform's compute capability, and its memory capacity and performance characteristics. To address this architectural customization problem, an automatic design space exploration (DSE) framework using a mixed-integer geometric programming (MIGP) approach is presented. Given the set of DNN models to be accelerated and a generic description of the target platform's compute and memory capabilities as input, the proposed framework automatically customizes an architectural template for the platform-model combination and produces the associated I/O schedule to maximize its end-to-end performance. By formulating DNN inference as a multi-level loop tiling problem, the proposed framework first customizes an accelerator template that consists of a parameterizable array architecture with SIMD execution cores and a customizable memory hierarchy using a MIGP to maximize the expected resource utilization. Subsequently, a second MIGP is used to schedule memory and compute operations as tiles to improve on-chip data reuse and memory bandwidth utilization. Experimental results from a wide range of neural network models and FPGA platform combinations show that the proposed scheme is able to produce accelerators with performance comparable to the state-of-the-art. The proposed DSE framework and the resulting hardware/software generator are available as an open-source package called AGNA with the hope that it may facilitate vendor-agnostic DNN accelerator development from the research community in the future. |
| Persistent Identifier | http://hdl.handle.net/10722/355680 |
| ISBN | |
| ISI Accession Number ID |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Ding, Yuhao | - |
| dc.contributor.author | Wu, Jiajun | - |
| dc.contributor.author | Gao, Yizhao | - |
| dc.contributor.author | Wang, Maolin | - |
| dc.contributor.author | So, Hayden Kwok-Hay | - |
| dc.date.accessioned | 2025-05-05T00:35:17Z | - |
| dc.date.available | 2025-05-05T00:35:17Z | - |
| dc.date.issued | 2023-07-01 | - |
| dc.identifier.isbn | 9798350312058 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/355680 | - |
| dc.description.abstract | Although there are distinct power-performance advantages in customizing an accelerator for a specific combination of FPGA platform and neural network model, developing such highly customized accelerators is a challenging task due to the massive design space spans from the range of network models to be accelerated, the target platform's compute capability, and its memory capacity and performance characteristics. To address this architectural customization problem, an automatic design space exploration (DSE) framework using a mixed-integer geometric programming (MIGP) approach is presented. Given the set of DNN models to be accelerated and a generic description of the target platform's compute and memory capabilities as input, the proposed framework automatically customizes an architectural template for the platform-model combination and produces the associated I/O schedule to maximize its end-to-end performance. By formulating DNN inference as a multi-level loop tiling problem, the proposed framework first customizes an accelerator template that consists of a parameterizable array architecture with SIMD execution cores and a customizable memory hierarchy using a MIGP to maximize the expected resource utilization. Subsequently, a second MIGP is used to schedule memory and compute operations as tiles to improve on-chip data reuse and memory bandwidth utilization. Experimental results from a wide range of neural network models and FPGA platform combinations show that the proposed scheme is able to produce accelerators with performance comparable to the state-of-the-art. The proposed DSE framework and the resulting hardware/software generator are available as an open-source package called AGNA with the hope that it may facilitate vendor-agnostic DNN accelerator development from the research community in the future. | - |
| dc.language | eng | - |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | - |
| dc.relation.ispartof | IEEE International Symposium on Field-Programmable Custom Computing Machines (08/05/2023-11/05/2023, Marina Del Rey, CA) | - |
| dc.subject | FPGA | - |
| dc.subject | hardware generation | - |
| dc.subject | machine learning | - |
| dc.title | Model-Platform Optimized Deep Neural Network Accelerator Generation through Mixed-Integer Geometric Programming | - |
| dc.type | Conference_Paper | - |
| dc.description.nature | preprint | - |
| dc.identifier.doi | 10.1109/FCCM57271.2023.00018 | - |
| dc.identifier.scopus | eid_2-s2.0-85165963734 | - |
| dc.identifier.spage | 83 | - |
| dc.identifier.epage | 93 | - |
| dc.identifier.isi | WOS:001032713500009 | - |
