File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1137/070707312
- Scopus: eid_2-s2.0-84981228964
- WOS: WOS:000278101100004
Supplementary
- Citations:
- Appears in Collections:
Article: Classification via minimum incremental coding length
Title | Classification via minimum incremental coding length |
---|---|
Authors | |
Keywords | Classification Lossy data coding Maximum a posteriori Regularization Regularized discriminant analysis |
Issue Date | 2009 |
Citation | SIAM Journal on Imaging Sciences, 2009, v. 2, n. 2, p. 367-395 How to Cite? |
Abstract | We present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the minimum number of additional bits to code the test sample, subject to an allowable distortion. We demonstrate the asymptotic optimality of this criterion for Gaussian distributions and analyze its relationships to classical classifiers. The theoretical results clarify the connections between our approach and popular classifiers such as maximum a posteriori (MAP), regularized discriminant analysis (RDA), k-nearest neighbor (k-NN), and support vector machine (SVM), as well as unsupervised methods based on lossy coding. Our formulation induces several good effects on the resulting classifier. First, minimizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small sample setting. Second, compression provides a uniform means of handling classes of varying dimension. The new criterion and its kernel and local versions perform competitively on synthetic examples, as well as on real imagery data such as handwritten digits and face images. On these problems, the performance of our simple classifier approaches the best reported results, without using domainspecific information. All MATLAB code and classification results are publicly available for peer evaluation at http://perception.csl.uiuc.edu/coding/home.htm. |
Persistent Identifier | http://hdl.handle.net/10722/327112 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Wright, John | - |
dc.contributor.author | Ma, Yi | - |
dc.contributor.author | Tao, Yangyu | - |
dc.contributor.author | Lin, Zhouchen | - |
dc.contributor.author | Shum, Heung Yeung | - |
dc.date.accessioned | 2023-03-31T05:28:53Z | - |
dc.date.available | 2023-03-31T05:28:53Z | - |
dc.date.issued | 2009 | - |
dc.identifier.citation | SIAM Journal on Imaging Sciences, 2009, v. 2, n. 2, p. 367-395 | - |
dc.identifier.uri | http://hdl.handle.net/10722/327112 | - |
dc.description.abstract | We present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the minimum number of additional bits to code the test sample, subject to an allowable distortion. We demonstrate the asymptotic optimality of this criterion for Gaussian distributions and analyze its relationships to classical classifiers. The theoretical results clarify the connections between our approach and popular classifiers such as maximum a posteriori (MAP), regularized discriminant analysis (RDA), k-nearest neighbor (k-NN), and support vector machine (SVM), as well as unsupervised methods based on lossy coding. Our formulation induces several good effects on the resulting classifier. First, minimizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small sample setting. Second, compression provides a uniform means of handling classes of varying dimension. The new criterion and its kernel and local versions perform competitively on synthetic examples, as well as on real imagery data such as handwritten digits and face images. On these problems, the performance of our simple classifier approaches the best reported results, without using domainspecific information. All MATLAB code and classification results are publicly available for peer evaluation at http://perception.csl.uiuc.edu/coding/home.htm. | - |
dc.language | eng | - |
dc.relation.ispartof | SIAM Journal on Imaging Sciences | - |
dc.subject | Classification | - |
dc.subject | Lossy data coding | - |
dc.subject | Maximum a posteriori | - |
dc.subject | Regularization | - |
dc.subject | Regularized discriminant analysis | - |
dc.title | Classification via minimum incremental coding length | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1137/070707312 | - |
dc.identifier.scopus | eid_2-s2.0-84981228964 | - |
dc.identifier.volume | 2 | - |
dc.identifier.issue | 2 | - |
dc.identifier.spage | 367 | - |
dc.identifier.epage | 395 | - |
dc.identifier.eissn | 1936-4954 | - |
dc.identifier.isi | WOS:000278101100004 | - |