Information Bottleneck for Gaussian Variables

Abstract

The problem of extracting the relevant aspects of data was ad- dressed through the information bottleneck (IB) method, by (soft) clustering one variable while preserving information about another - relevance - variable. An interesting question addressed in the current work is the extension of these ideas to obtain continuous representations that preserve relevant information, rather than dis- crete clusters. We give a formal deflnition of the general continuous IB problem and obtain an analytic solution for the optimal repre- sentation for the important case of multivariate Gaussian variables. The obtained optimal representation is a noisy linear projection to eigenvectors of the normalized correlation matrix §xjy§¡1 x , which is also the basis obtained in Canonical Correlation Analysis. How- ever, in Gaussian IB, the compression tradeofi parameter uniquely determines the dimension, as well as the scale of each eigenvector. This introduces a novel interpretation where solutions of difierent ranks lie on a continuum parametrized by the compression level. Our analysis also provides an analytic expression for the optimal tradeofi - the information curve - in terms of the eigenvalue spec- trum.

Cite

Text

Chechik et al. "Information Bottleneck for Gaussian Variables." Neural Information Processing Systems, 2003.

Markdown

[Chechik et al. "Information Bottleneck for Gaussian Variables." Neural Information Processing Systems, 2003.](https://mlanthology.org/neurips/2003/chechik2003neurips-information/)

BibTeX

@inproceedings{chechik2003neurips-information,
  title     = {{Information Bottleneck for Gaussian Variables}},
  author    = {Chechik, Gal and Globerson, Amir and Tishby, Naftali and Weiss, Yair},
  booktitle = {Neural Information Processing Systems},
  year      = {2003},
  pages     = {1213-1220},
  url       = {https://mlanthology.org/neurips/2003/chechik2003neurips-information/}
}