InfoNCE Is Variational Inference in a Recognition Parameterised Model

Abstract

Here, we develop a new class of Bayesian latent variable model, the recognition parameterised model (RPM). RPMs have an implicit likelihood, which is defined in terms of the recognition model. Therefore, it is not possible to do traditional "generation" with RPMs. Instead, RPMs are designed to learn good latent representations of data (in modern parlance, they solve a self-supervised learning task). Indeed, the RPM implicit likelihood is specifically designed so that it drops out of the VI objective, the ELBO. That allows us to learn an RPM without a "reconstruction" step, which is believed to be at the root of poor latent representations learned by VAEs. Indeed, in a very specific setting where we learn the optimal prior, the RPM ELBO becomes equal to the mutual information (MI; up to a constant), establishing a connection to pre-existing self-supervised learning methods such as InfoNCE.

Cite

Text

Aitchison and Ganev. "InfoNCE Is Variational Inference in a Recognition Parameterised Model." Transactions on Machine Learning Research, 2024.

Markdown

[Aitchison and Ganev. "InfoNCE Is Variational Inference in a Recognition Parameterised Model." Transactions on Machine Learning Research, 2024.](https://mlanthology.org/tmlr/2024/aitchison2024tmlr-infonce/)

BibTeX

@article{aitchison2024tmlr-infonce,
  title     = {{InfoNCE Is Variational Inference in a Recognition Parameterised Model}},
  author    = {Aitchison, Laurence and Ganev, Stoil Krasimirov},
  journal   = {Transactions on Machine Learning Research},
  year      = {2024},
  url       = {https://mlanthology.org/tmlr/2024/aitchison2024tmlr-infonce/}
}