Filtering Variational Objectives

Abstract

When used as a surrogate objective for maximum likelihood estimation in latent variable models, the evidence lower bound (ELBO) produces state-of-the-art results. Inspired by this, we consider the extension of the ELBO to a family of lower bounds defined by a particle filter's estimator of the marginal likelihood, the filtering variational objectives (FIVOs). FIVOs take the same arguments as the ELBO, but can exploit a model's sequential structure to form tighter bounds. We present results that relate the tightness of FIVO's bound to the variance of the particle filter's estimator by considering the generic case of bounds defined as log-transformed likelihood estimators. Experimentally, we show that training with FIVO results in substantial improvements over training the same model architecture with the ELBO on sequential data.

Cite

Text

Maddison et al. "Filtering Variational Objectives." Neural Information Processing Systems, 2017.

Markdown

[Maddison et al. "Filtering Variational Objectives." Neural Information Processing Systems, 2017.](https://mlanthology.org/neurips/2017/maddison2017neurips-filtering/)

BibTeX

@inproceedings{maddison2017neurips-filtering,
  title     = {{Filtering Variational Objectives}},
  author    = {Maddison, Chris J and Lawson, John and Tucker, George and Heess, Nicolas and Norouzi, Mohammad and Mnih, Andriy and Doucet, Arnaud and Teh, Yee},
  booktitle = {Neural Information Processing Systems},
  year      = {2017},
  pages     = {6573-6583},
  url       = {https://mlanthology.org/neurips/2017/maddison2017neurips-filtering/}
}