File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction
Title | ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction |
---|---|
Authors | |
Keywords | linear discriminative representation multi-channel convolution rate reduction sparsity and invariance trade-off white-box deep network |
Issue Date | 2022 |
Citation | Journal of Machine Learning Research, 2022, v. 23 How to Cite? |
Abstract | This work attempts to provide a plausible theoretical framework that aims to interpret modern deep (convolutional) networks from the principles of data compression and discriminative representation. We argue that for high-dimensional multi-class data, the optimal linear discriminative representation maximizes the coding rate difference between the whole dataset and the average of all the subsets. We show that the basic iterative gradient ascent scheme for optimizing the rate reduction objective naturally leads to a multi-layer deep network, named ReduNet, which shares common characteristics of modern deep networks. The deep layered architectures, linear and nonlinear operators, and even parameters of the network are all explicitly constructed layer-by-layer via forward propagation, although they are amenable to fine-tuning via back propagation. All components of so-obtained \white-box"network have precise optimization, statistical, and geometric interpretation. Moreover, all linear operators of the so-derived network naturally become multi-channel convolutions when we enforce classification to be rigorously shift-invariant. The derivation in the invariant setting suggests a trade-off between sparsity and invariance, and also indicates that such a deep convolution network is significantly more efficient to construct and learn in the spectral domain. Our preliminary simulations and experiments clearly verify the effectiveness of both the rate reduction objective and the associated ReduNet. |
Persistent Identifier | http://hdl.handle.net/10722/327784 |
ISSN | 2023 Impact Factor: 4.3 2023 SCImago Journal Rankings: 2.796 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chan, Kwan Ho Ryan | - |
dc.contributor.author | Yu, Yaodong | - |
dc.contributor.author | You, Chong | - |
dc.contributor.author | Qi, Haozhi | - |
dc.contributor.author | Wright, John | - |
dc.contributor.author | Ma, Yi | - |
dc.date.accessioned | 2023-05-08T02:26:47Z | - |
dc.date.available | 2023-05-08T02:26:47Z | - |
dc.date.issued | 2022 | - |
dc.identifier.citation | Journal of Machine Learning Research, 2022, v. 23 | - |
dc.identifier.issn | 1532-4435 | - |
dc.identifier.uri | http://hdl.handle.net/10722/327784 | - |
dc.description.abstract | This work attempts to provide a plausible theoretical framework that aims to interpret modern deep (convolutional) networks from the principles of data compression and discriminative representation. We argue that for high-dimensional multi-class data, the optimal linear discriminative representation maximizes the coding rate difference between the whole dataset and the average of all the subsets. We show that the basic iterative gradient ascent scheme for optimizing the rate reduction objective naturally leads to a multi-layer deep network, named ReduNet, which shares common characteristics of modern deep networks. The deep layered architectures, linear and nonlinear operators, and even parameters of the network are all explicitly constructed layer-by-layer via forward propagation, although they are amenable to fine-tuning via back propagation. All components of so-obtained \white-box"network have precise optimization, statistical, and geometric interpretation. Moreover, all linear operators of the so-derived network naturally become multi-channel convolutions when we enforce classification to be rigorously shift-invariant. The derivation in the invariant setting suggests a trade-off between sparsity and invariance, and also indicates that such a deep convolution network is significantly more efficient to construct and learn in the spectral domain. Our preliminary simulations and experiments clearly verify the effectiveness of both the rate reduction objective and the associated ReduNet. | - |
dc.language | eng | - |
dc.relation.ispartof | Journal of Machine Learning Research | - |
dc.subject | linear discriminative representation | - |
dc.subject | multi-channel convolution | - |
dc.subject | rate reduction | - |
dc.subject | sparsity and invariance trade-off | - |
dc.subject | white-box deep network | - |
dc.title | ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.scopus | eid_2-s2.0-85130327907 | - |
dc.identifier.volume | 23 | - |
dc.identifier.eissn | 1533-7928 | - |