Kullback-Leibler Proximal Variational Inference

Abstract

We propose a new variational inference method based on the Kullback-Leibler (KL) proximal term. We make two contributions towards improving efficiency of variational inference. Firstly, we derive a KL proximal-point algorithm and show its equivalence to gradient descent with natural gradient in stochastic variational inference. Secondly, we use the proximal framework to derive efficient variational algorithms for non-conjugate models. We propose a splitting procedure to separate non-conjugate terms from conjugate ones. We then linearize the non-conjugate terms and show that the resulting subproblem admits a closed-form solution. Overall, our approach converts a non-conjugate model to subproblems that involve inference in well-known conjugate models. We apply our method to many models and derive generalizations for non-conjugate exponential family. Applications to real-world datasets show that our proposed algorithms are easy to implement, fast to converge, perform well, and reduce computations.

Cite

Text

Khan et al. "Kullback-Leibler Proximal Variational Inference." Neural Information Processing Systems, 2015.

Markdown

[Khan et al. "Kullback-Leibler Proximal Variational Inference." Neural Information Processing Systems, 2015.](https://mlanthology.org/neurips/2015/khan2015neurips-kullbackleibler/)

BibTeX

@inproceedings{khan2015neurips-kullbackleibler,
  title     = {{Kullback-Leibler Proximal Variational Inference}},
  author    = {Khan, Mohammad Emtiyaz and Baque, Pierre and Fleuret, François and Fua, Pascal},
  booktitle = {Neural Information Processing Systems},
  year      = {2015},
  pages     = {3402-3410},
  url       = {https://mlanthology.org/neurips/2015/khan2015neurips-kullbackleibler/}
}