File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TNNLS.2022.3190607
- Find via
Supplementary
-
Citations:
- Appears in Collections:
Article: FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations
Title | FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations |
---|---|
Authors | |
Issue Date | 21-Oct-2022 |
Publisher | Institute of Electrical and Electronics Engineers |
Citation | IEEE Transactions on Neural Networks and Learning Systems, 2022, v. 13 How to Cite? |
Abstract | Learning low-bitwidth convolutional neural networks (CNNs) is challenging because performance may drop significantly after quantization. Prior arts often quantize the network weights by carefully tuning hyperparameters such as nonuniform stepsize and layerwise bitwidths, which are complicated since the full-and low-precision representations have large discrepancies. This work presents a novel quantization pipeline, named frequency-aware transformation (FAT), that features important benefits: 1) instead of designing complicated quantizers, FAT learns to transform network weights in the frequency domain to remove redundant information before quantization, making them amenable to training in low bitwidth with simple quantizers; 2) FAT readily embeds CNNs in low bitwidths using standard quantizers without tedious hyperparameter tuning and theoretical analyses show that FAT minimizes the quantization errors in both uniform and nonuniform quantizations; and 3) FAT can be easily plugged into various CNN architectures. Using FAT with a simple uniform/logarithmic quantizer can achieve the state-of-the-art performance in different bitwidths on various model architectures. Consequently, FAT serves to provide a novel frequency-based perspective for model quantization. |
Persistent Identifier | http://hdl.handle.net/10722/339470 |
ISSN | 2023 Impact Factor: 10.2 2023 SCImago Journal Rankings: 4.170 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Tao, Chaofan | - |
dc.contributor.author | Lin, Rui | - |
dc.contributor.author | Chen, Quan | - |
dc.contributor.author | Zhang, Zhaoyang | - |
dc.contributor.author | Luo, Ping | - |
dc.contributor.author | Wong, Ngai | - |
dc.date.accessioned | 2024-03-11T10:36:54Z | - |
dc.date.available | 2024-03-11T10:36:54Z | - |
dc.date.issued | 2022-10-21 | - |
dc.identifier.citation | IEEE Transactions on Neural Networks and Learning Systems, 2022, v. 13 | - |
dc.identifier.issn | 2162-237X | - |
dc.identifier.uri | http://hdl.handle.net/10722/339470 | - |
dc.description.abstract | <p>Learning low-bitwidth convolutional neural networks (CNNs) is challenging because performance may drop significantly after quantization. Prior arts often quantize the network weights by carefully tuning hyperparameters such as nonuniform stepsize and layerwise bitwidths, which are complicated since the full-and low-precision representations have large discrepancies. This work presents a novel quantization pipeline, named frequency-aware transformation (FAT), that features important benefits: 1) instead of designing complicated quantizers, FAT learns to transform network weights in the frequency domain to remove redundant information before quantization, making them amenable to training in low bitwidth with simple quantizers; 2) FAT readily embeds CNNs in low bitwidths using standard quantizers without tedious hyperparameter tuning and theoretical analyses show that FAT minimizes the quantization errors in both uniform and nonuniform quantizations; and 3) FAT can be easily plugged into various CNN architectures. Using FAT with a simple uniform/logarithmic quantizer can achieve the state-of-the-art performance in different bitwidths on various model architectures. Consequently, FAT serves to provide a novel frequency-based perspective for model quantization.<br></p> | - |
dc.language | eng | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.relation.ispartof | IEEE Transactions on Neural Networks and Learning Systems | - |
dc.title | FAT: Frequency-Aware Transformation for Bridging Full-Precision and Low-Precision Deep Representations | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/TNNLS.2022.3190607 | - |
dc.identifier.volume | 13 | - |
dc.identifier.eissn | 2162-2388 | - |
dc.identifier.issnl | 2162-237X | - |