Variational Inference via Wasserstein Gradient Flows

Abstract

Along with Markov chain Monte Carlo (MCMC) methods, variational inference (VI) has emerged as a central computational approach to large-scale Bayesian inference. Rather than sampling from the true posterior $\pi$, VI aims at producing a simple but effective approximation $\hat \pi$ to $\pi$ for which summary statistics are easy to compute. However, unlike the well-studied MCMC methodology, algorithmic guarantees for VI are still relatively less well-understood. In this work, we propose principled methods for VI, in which $\hat \pi$ is taken to be a Gaussian or a mixture of Gaussians, which rest upon the theory of gradient flows on the Bures--Wasserstein space of Gaussian measures. Akin to MCMC, it comes with strong theoretical guarantees when $\pi$ is log-concave.

Cite

Text

Lambert et al. "Variational Inference via Wasserstein Gradient Flows." Neural Information Processing Systems, 2022.

Markdown

[Lambert et al. "Variational Inference via Wasserstein Gradient Flows." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/lambert2022neurips-variational/)

BibTeX

@inproceedings{lambert2022neurips-variational,
  title     = {{Variational Inference via Wasserstein Gradient Flows}},
  author    = {Lambert, Marc and Chewi, Sinho and Bach, Francis R. and Bonnabel, Silvère and Rigollet, Philippe},
  booktitle = {Neural Information Processing Systems},
  year      = {2022},
  url       = {https://mlanthology.org/neurips/2022/lambert2022neurips-variational/}
}