File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.3390/e21050456
- Scopus: eid_2-s2.0-85066605153
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: Utilizing information bottleneck to evaluate the capability of deep neural networks for image classification
Title | Utilizing information bottleneck to evaluate the capability of deep neural networks for image classification |
---|---|
Authors | |
Keywords | Image classification Information bottleneck Mutual information Neural networks |
Issue Date | 2019 |
Citation | Entropy, 2019, v. 21, n. 5, article no. 456 How to Cite? |
Abstract | Inspired by the pioneering work of the information bottleneck (IB) principle for Deep Neural Networks' (DNNs) analysis, we thoroughly study the relationship among the model accuracy, I(X; T) and I(T;Y), where I(X; T) and I(T;Y) are the mutual information of DNN's output T with input X and label Y. Then, we design an information plane-based framework to evaluate the capability of DNNs (including CNNs) for image classification. Instead of each hidden layer's output, our framework focuses on the model output T. We successfully apply our framework to many application scenarios arising in deep learning and image classification problems, such as image classification with unbalanced data distribution, model selection, and transfer learning. The experimental results verify the effectiveness of the information plane-based framework: Our framework may facilitate a quick model selection and determine the number of samples needed for each class in the unbalanced classification problem. Furthermore, the framework explains the efficiency of transfer learning in the deep learning area. |
Persistent Identifier | http://hdl.handle.net/10722/345247 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Cheng, Hao | - |
dc.contributor.author | Lian, Dongze | - |
dc.contributor.author | Gao, Shenghua | - |
dc.contributor.author | Geng, Yanlin | - |
dc.date.accessioned | 2024-08-15T09:26:09Z | - |
dc.date.available | 2024-08-15T09:26:09Z | - |
dc.date.issued | 2019 | - |
dc.identifier.citation | Entropy, 2019, v. 21, n. 5, article no. 456 | - |
dc.identifier.uri | http://hdl.handle.net/10722/345247 | - |
dc.description.abstract | Inspired by the pioneering work of the information bottleneck (IB) principle for Deep Neural Networks' (DNNs) analysis, we thoroughly study the relationship among the model accuracy, I(X; T) and I(T;Y), where I(X; T) and I(T;Y) are the mutual information of DNN's output T with input X and label Y. Then, we design an information plane-based framework to evaluate the capability of DNNs (including CNNs) for image classification. Instead of each hidden layer's output, our framework focuses on the model output T. We successfully apply our framework to many application scenarios arising in deep learning and image classification problems, such as image classification with unbalanced data distribution, model selection, and transfer learning. The experimental results verify the effectiveness of the information plane-based framework: Our framework may facilitate a quick model selection and determine the number of samples needed for each class in the unbalanced classification problem. Furthermore, the framework explains the efficiency of transfer learning in the deep learning area. | - |
dc.language | eng | - |
dc.relation.ispartof | Entropy | - |
dc.subject | Image classification | - |
dc.subject | Information bottleneck | - |
dc.subject | Mutual information | - |
dc.subject | Neural networks | - |
dc.title | Utilizing information bottleneck to evaluate the capability of deep neural networks for image classification | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.3390/e21050456 | - |
dc.identifier.scopus | eid_2-s2.0-85066605153 | - |
dc.identifier.volume | 21 | - |
dc.identifier.issue | 5 | - |
dc.identifier.spage | article no. 456 | - |
dc.identifier.epage | article no. 456 | - |
dc.identifier.eissn | 1099-4300 | - |