A Generalised Linear Model Framework for Β-Variational Autoencoders Based on Exponential Dispersion Families

Abstract

Although variational autoencoders (VAE) are successfully used to obtain meaningful low-dimensional representations for high-dimensional data, the characterization of critical points of the loss function for general observation models is not fully understood. We introduce a theoretical framework that is based on a connection between β-VAE and generalized linear models (GLM). The equality between the activation function of a β-VAE and the inverse of the link function of a GLM enables us to provide a systematic generalization of the loss analysis for β-VAE based on the assumption that the observation model distribution belongs to an exponential dispersion family (EDF). As a result, we can initialize β-VAE nets by maximum likelihood estimates (MLE) that enhance the training performance on both synthetic and real world data sets. As a further consequence, we analytically describe the auto-pruning property inherent in the β-VAE objective and reason for posterior collapse.

Cite

Text

Sicks et al. "A Generalised Linear Model Framework for Β-Variational Autoencoders Based on Exponential Dispersion Families." Journal of Machine Learning Research, 2021.

Markdown

[Sicks et al. "A Generalised Linear Model Framework for Β-Variational Autoencoders Based on Exponential Dispersion Families." Journal of Machine Learning Research, 2021.](https://mlanthology.org/jmlr/2021/sicks2021jmlr-generalised/)

BibTeX

@article{sicks2021jmlr-generalised,
  title     = {{A Generalised Linear Model Framework for Β-Variational Autoencoders Based on Exponential Dispersion Families}},
  author    = {Sicks, Robert and Korn, Ralf and Schwaar, Stefanie},
  journal   = {Journal of Machine Learning Research},
  year      = {2021},
  pages     = {1-41},
  volume    = {22},
  url       = {https://mlanthology.org/jmlr/2021/sicks2021jmlr-generalised/}
}