Classification via Minimum Incremental Coding Length (MICL)
Abstract
We present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the min- imum number of additional bits to code the test sample, subject to an allowable distortion. We prove asymptotic optimality of this criterion for Gaussian data and analyze its relationships to classical classifiers. Theoretical results provide new insights into relationships among popular classifiers such as MAP and RDA, as well as unsupervised clustering methods based on lossy compression [13]. Mini- mizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small-sample setting. Compression also provides a uniform means of handling classes of varying dimension. This simple classi- fication criterion and its kernel and local versions perform competitively against existing classifiers on both synthetic examples and real imagery data such as hand- written digits and human faces, without requiring domain-specific information.
Cite
Text
Wright et al. "Classification via Minimum Incremental Coding Length (MICL)." Neural Information Processing Systems, 2007.Markdown
[Wright et al. "Classification via Minimum Incremental Coding Length (MICL)." Neural Information Processing Systems, 2007.](https://mlanthology.org/neurips/2007/wright2007neurips-classification/)BibTeX
@inproceedings{wright2007neurips-classification,
title = {{Classification via Minimum Incremental Coding Length (MICL)}},
author = {Wright, John and Tao, Yangyu and Lin, Zhouchen and Ma, Yi and Shum, Heung-yeung},
booktitle = {Neural Information Processing Systems},
year = {2007},
pages = {1633-1640},
url = {https://mlanthology.org/neurips/2007/wright2007neurips-classification/}
}