Auto-Encoding Total Correlation Explanation

Abstract

Advances in unsupervised learning enable reconstruction and generation of samples from complex distributions, but this success is marred by the inscrutability of the representations learned. We propose an information-theoretic approach to characterizing disentanglement and dependence in representation learning using multivariate mutual information, also called total correlation. The principle of Total Cor-relation Ex-planation (CorEx) has motivated successful unsupervised learning applications across a variety of domains but under some restrictive assumptions. Here we relax those restrictions by introducing a flexible variational lower bound to CorEx. Surprisingly, we find this lower bound is equivalent to the one in variational autoencoders (VAE) under certain conditions. This information-theoretic view of VAE deepens our understanding of hierarchical VAE and motivates a new algorithm, AnchorVAE, that makes latent codes more interpretable through information maximization and enables generation of richer and more realistic samples.

Cite

Text

Gao et al. "Auto-Encoding Total Correlation Explanation." Artificial Intelligence and Statistics, 2019.

Markdown

[Gao et al. "Auto-Encoding Total Correlation Explanation." Artificial Intelligence and Statistics, 2019.](https://mlanthology.org/aistats/2019/gao2019aistats-autoencoding/)

BibTeX

@inproceedings{gao2019aistats-autoencoding,
  title     = {{Auto-Encoding Total Correlation Explanation}},
  author    = {Gao, Shuyang and Brekelmans, Rob and Steeg, Greg Ver and Galstyan, Aram},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2019},
  pages     = {1157-1166},
  volume    = {89},
  url       = {https://mlanthology.org/aistats/2019/gao2019aistats-autoencoding/}
}