Relevant Sparse Codes with Variational Information Bottleneck

Abstract

In many applications, it is desirable to extract only the relevant aspects of data. A principled way to do this is the information bottleneck (IB) method, where one seeks a code that maximises information about a relevance variable, Y, while constraining the information encoded about the original data, X. Unfortunately however, the IB method is computationally demanding when data are high-dimensional and/or non-gaussian. Here we propose an approximate variational scheme for maximising a lower bound on the IB objective, analogous to variational EM. Using this method, we derive an IB algorithm to recover features that are both relevant and sparse. Finally, we demonstrate how kernelised versions of the algorithm can be used to address a broad range of problems with non-linear relation between X and Y.

Cite

Text

Chalk et al. "Relevant Sparse Codes with Variational Information Bottleneck." Neural Information Processing Systems, 2016.

Markdown

[Chalk et al. "Relevant Sparse Codes with Variational Information Bottleneck." Neural Information Processing Systems, 2016.](https://mlanthology.org/neurips/2016/chalk2016neurips-relevant/)

BibTeX

@inproceedings{chalk2016neurips-relevant,
  title     = {{Relevant Sparse Codes with Variational Information Bottleneck}},
  author    = {Chalk, Matthew and Marre, Olivier and Tkacik, Gasper},
  booktitle = {Neural Information Processing Systems},
  year      = {2016},
  pages     = {1957-1965},
  url       = {https://mlanthology.org/neurips/2016/chalk2016neurips-relevant/}
}