File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Model-Platform Optimized Deep Neural Network Accelerator Generation through Mixed-Integer Geometric Programming

TitleModel-Platform Optimized Deep Neural Network Accelerator Generation through Mixed-Integer Geometric Programming
Authors
KeywordsFPGA
hardware generation
machine learning
Issue Date1-Jul-2023
PublisherInstitute of Electrical and Electronics Engineers Inc.
AbstractAlthough there are distinct power-performance advantages in customizing an accelerator for a specific combination of FPGA platform and neural network model, developing such highly customized accelerators is a challenging task due to the massive design space spans from the range of network models to be accelerated, the target platform's compute capability, and its memory capacity and performance characteristics. To address this architectural customization problem, an automatic design space exploration (DSE) framework using a mixed-integer geometric programming (MIGP) approach is presented. Given the set of DNN models to be accelerated and a generic description of the target platform's compute and memory capabilities as input, the proposed framework automatically customizes an architectural template for the platform-model combination and produces the associated I/O schedule to maximize its end-to-end performance. By formulating DNN inference as a multi-level loop tiling problem, the proposed framework first customizes an accelerator template that consists of a parameterizable array architecture with SIMD execution cores and a customizable memory hierarchy using a MIGP to maximize the expected resource utilization. Subsequently, a second MIGP is used to schedule memory and compute operations as tiles to improve on-chip data reuse and memory bandwidth utilization. Experimental results from a wide range of neural network models and FPGA platform combinations show that the proposed scheme is able to produce accelerators with performance comparable to the state-of-the-art. The proposed DSE framework and the resulting hardware/software generator are available as an open-source package called AGNA with the hope that it may facilitate vendor-agnostic DNN accelerator development from the research community in the future.
Persistent Identifierhttp://hdl.handle.net/10722/355680
ISBN
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorDing, Yuhao-
dc.contributor.authorWu, Jiajun-
dc.contributor.authorGao, Yizhao-
dc.contributor.authorWang, Maolin-
dc.contributor.authorSo, Hayden Kwok-Hay-
dc.date.accessioned2025-05-05T00:35:17Z-
dc.date.available2025-05-05T00:35:17Z-
dc.date.issued2023-07-01-
dc.identifier.isbn9798350312058-
dc.identifier.urihttp://hdl.handle.net/10722/355680-
dc.description.abstractAlthough there are distinct power-performance advantages in customizing an accelerator for a specific combination of FPGA platform and neural network model, developing such highly customized accelerators is a challenging task due to the massive design space spans from the range of network models to be accelerated, the target platform's compute capability, and its memory capacity and performance characteristics. To address this architectural customization problem, an automatic design space exploration (DSE) framework using a mixed-integer geometric programming (MIGP) approach is presented. Given the set of DNN models to be accelerated and a generic description of the target platform's compute and memory capabilities as input, the proposed framework automatically customizes an architectural template for the platform-model combination and produces the associated I/O schedule to maximize its end-to-end performance. By formulating DNN inference as a multi-level loop tiling problem, the proposed framework first customizes an accelerator template that consists of a parameterizable array architecture with SIMD execution cores and a customizable memory hierarchy using a MIGP to maximize the expected resource utilization. Subsequently, a second MIGP is used to schedule memory and compute operations as tiles to improve on-chip data reuse and memory bandwidth utilization. Experimental results from a wide range of neural network models and FPGA platform combinations show that the proposed scheme is able to produce accelerators with performance comparable to the state-of-the-art. The proposed DSE framework and the resulting hardware/software generator are available as an open-source package called AGNA with the hope that it may facilitate vendor-agnostic DNN accelerator development from the research community in the future.-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers Inc.-
dc.relation.ispartofIEEE International Symposium on Field-Programmable Custom Computing Machines (08/05/2023-11/05/2023, Marina Del Rey, CA)-
dc.subjectFPGA-
dc.subjecthardware generation-
dc.subjectmachine learning-
dc.titleModel-Platform Optimized Deep Neural Network Accelerator Generation through Mixed-Integer Geometric Programming-
dc.typeConference_Paper-
dc.description.naturepreprint-
dc.identifier.doi10.1109/FCCM57271.2023.00018-
dc.identifier.scopuseid_2-s2.0-85165963734-
dc.identifier.spage83-
dc.identifier.epage93-
dc.identifier.isiWOS:001032713500009-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats