F-Divergence Variational Inference
Abstract
This paper introduces the f-divergence variational inference (f-VI) that generalizes variational inference to all f-divergences. Initiated from minimizing a crafty surrogate f-divergence that shares the statistical consistency with the f-divergence, the f-VI framework not only unifies a number of existing VI methods, e.g. Kullback–Leibler VI, Renyi's alpha-VI, and chi-VI, but offers a standardized toolkit for VI subject to arbitrary divergences from f-divergence family. A general f-variational bound is derived and provides a sandwich estimate of marginal likelihood (or evidence). The development of the f-VI unfolds with a stochastic optimization scheme that utilizes the reparameterization trick, importance weighting and Monte Carlo approximation; a mean-field approximation scheme that generalizes the well-known coordinate ascent variational inference (CAVI) is also proposed for f-VI. Empirical examples, including variational autoencoders and Bayesian neural networks, are provided to demonstrate the effectiveness and the wide applicability of f-VI.
Cite
Text
Wan et al. "F-Divergence Variational Inference." Neural Information Processing Systems, 2020.Markdown
[Wan et al. "F-Divergence Variational Inference." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/wan2020neurips-fdivergence/)BibTeX
@inproceedings{wan2020neurips-fdivergence,
title = {{F-Divergence Variational Inference}},
author = {Wan, Neng and Li, Dapeng and Hovakimyan, Naira},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/wan2020neurips-fdivergence/}
}