File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TNNLS.2023.3333562
- Scopus: eid_2-s2.0-85179062481
- WOS: WOS:001121219100001
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Lite It Fly: An All-Deformable-Butterfly Network
Title | Lite It Fly: An All-Deformable-Butterfly Network |
---|---|
Authors | |
Keywords | Convolutional neural network (CNN) Convolutional neural networks Hardware Kernel Learning systems Matrix decomposition model compression Shape Sparse matrices structured sparse matrix |
Issue Date | 28-Nov-2023 |
Publisher | Institute of Electrical and Electronics Engineers |
Citation | IEEE Transactions on Neural Networks and Learning Systems, 2023, p. 1-6 How to Cite? |
Abstract | Most deep neural networks (DNNs) consist fundamentally of convolutional and/or fully connected layers, wherein the linear transform can be cast as the product between a filter matrix and a data matrix obtained by arranging feature tensors into columns. Recently proposed deformable butterfly (DeBut) decomposes the filter matrix into generalized, butterfly-like factors, thus achieving network compression orthogonal to the traditional ways of pruning or low-rank decomposition. This work reveals an intimate link between DeBut and a systematic hierarchy of depthwise and pointwise convolutions, which explains the empirically good performance of DeBut layers. By developing an automated DeBut chain generator, we show for the first time the viability of homogenizing a DNN into all DeBut layers, thus achieving extreme sparsity and compression. Various examples and hardware benchmarks verify the advantages of All-DeBut networks. In particular, we show it is possible to compress a PointNet to < 5% parameters with < 5% accuracy drop, a record not achievable by other compression schemes. |
Persistent Identifier | http://hdl.handle.net/10722/339468 |
ISSN | 2023 Impact Factor: 10.2 2023 SCImago Journal Rankings: 4.170 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lin, Rui | - |
dc.contributor.author | Li, Jason Chun Lok | - |
dc.contributor.author | Zhou, Jiajun | - |
dc.contributor.author | Huang, Binxiao | - |
dc.contributor.author | Ran, Jie | - |
dc.contributor.author | Wong, Ngai | - |
dc.date.accessioned | 2024-03-11T10:36:53Z | - |
dc.date.available | 2024-03-11T10:36:53Z | - |
dc.date.issued | 2023-11-28 | - |
dc.identifier.citation | IEEE Transactions on Neural Networks and Learning Systems, 2023, p. 1-6 | - |
dc.identifier.issn | 2162-237X | - |
dc.identifier.uri | http://hdl.handle.net/10722/339468 | - |
dc.description.abstract | <p>Most deep neural networks (DNNs) consist fundamentally of convolutional and/or fully connected layers, wherein the linear transform can be cast as the product between a filter matrix and a data matrix obtained by arranging feature tensors into columns. Recently proposed deformable butterfly (DeBut) decomposes the filter matrix into generalized, butterfly-like factors, thus achieving network compression orthogonal to the traditional ways of pruning or low-rank decomposition. This work reveals an intimate link between DeBut and a systematic hierarchy of depthwise and pointwise convolutions, which explains the empirically good performance of DeBut layers. By developing an automated DeBut chain generator, we show for the first time the viability of homogenizing a DNN into all DeBut layers, thus achieving extreme sparsity and compression. Various examples and hardware benchmarks verify the advantages of All-DeBut networks. In particular, we show it is possible to compress a PointNet to < 5% parameters with < 5% accuracy drop, a record not achievable by other compression schemes.<br></p> | - |
dc.language | eng | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.relation.ispartof | IEEE Transactions on Neural Networks and Learning Systems | - |
dc.subject | Convolutional neural network (CNN) | - |
dc.subject | Convolutional neural networks | - |
dc.subject | Hardware | - |
dc.subject | Kernel | - |
dc.subject | Learning systems | - |
dc.subject | Matrix decomposition | - |
dc.subject | model compression | - |
dc.subject | Shape | - |
dc.subject | Sparse matrices | - |
dc.subject | structured sparse matrix | - |
dc.title | Lite It Fly: An All-Deformable-Butterfly Network | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/TNNLS.2023.3333562 | - |
dc.identifier.scopus | eid_2-s2.0-85179062481 | - |
dc.identifier.spage | 1 | - |
dc.identifier.epage | 6 | - |
dc.identifier.eissn | 2162-2388 | - |
dc.identifier.isi | WOS:001121219100001 | - |
dc.identifier.issnl | 2162-237X | - |