File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Classification via minimum incremental coding length (MICL)
Title | Classification via minimum incremental coding length (MICL) |
---|---|
Authors | |
Issue Date | 2008 |
Citation | Advances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conference, 2008 How to Cite? |
Abstract | We present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the minimum number of additional bits to code the test sample, subject to an allowable distortion. We prove asymptotic optimality of this criterion for Gaussian data and analyze its relationships to classical classifiers. Theoretical results provide new insights into relationships among popular classifiers such as MAP and RDA, as well as unsupervised clustering methods based on lossy compression [13]. Minimizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small-sample setting. Compression also provides a uniform means of handling classes of varying dimension. This simple classification criterion and its kernel and local versions perform competitively against existing classifiers on both synthetic examples and real imagery data such as handwritten digits and human faces, without requiring domain-specific information. |
Persistent Identifier | http://hdl.handle.net/10722/326622 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Wright, John | - |
dc.contributor.author | Ma, Yi | - |
dc.contributor.author | Tao, Yangyu | - |
dc.contributor.author | Lin, Zhouchen | - |
dc.contributor.author | Shum, Heung Yeung | - |
dc.date.accessioned | 2023-03-31T05:25:18Z | - |
dc.date.available | 2023-03-31T05:25:18Z | - |
dc.date.issued | 2008 | - |
dc.identifier.citation | Advances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conference, 2008 | - |
dc.identifier.uri | http://hdl.handle.net/10722/326622 | - |
dc.description.abstract | We present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the minimum number of additional bits to code the test sample, subject to an allowable distortion. We prove asymptotic optimality of this criterion for Gaussian data and analyze its relationships to classical classifiers. Theoretical results provide new insights into relationships among popular classifiers such as MAP and RDA, as well as unsupervised clustering methods based on lossy compression [13]. Minimizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small-sample setting. Compression also provides a uniform means of handling classes of varying dimension. This simple classification criterion and its kernel and local versions perform competitively against existing classifiers on both synthetic examples and real imagery data such as handwritten digits and human faces, without requiring domain-specific information. | - |
dc.language | eng | - |
dc.relation.ispartof | Advances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conference | - |
dc.title | Classification via minimum incremental coding length (MICL) | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.scopus | eid_2-s2.0-85138462036 | - |