REBAR: Low-Variance, Unbiased Gradient Estimates for Discrete Latent Variable Models

Abstract

Learning in models with discrete latent variables is challenging due to high variance gradient estimators. Generally, approaches have relied on control variates to reduce the variance of the REINFORCE estimator. Recent work \citep{jang2016categorical, maddison2016concrete} has taken a different approach, introducing a continuous relaxation of discrete variables to produce low-variance, but biased, gradient estimates. In this work, we combine the two approaches through a novel control variate that produces low-variance, \emph{unbiased} gradient estimates. Then, we introduce a modification to the continuous relaxation and show that the tightness of the relaxation can be adapted online, removing it as a hyperparameter. We show state-of-the-art variance reduction on several benchmark generative modeling tasks, generally leading to faster convergence to a better final log-likelihood.

Cite

Text

Tucker et al. "REBAR: Low-Variance, Unbiased Gradient Estimates for Discrete Latent Variable Models." International Conference on Learning Representations, 2017.

Markdown

[Tucker et al. "REBAR: Low-Variance, Unbiased Gradient Estimates for Discrete Latent Variable Models." International Conference on Learning Representations, 2017.](https://mlanthology.org/iclr/2017/tucker2017iclr-rebar/)

BibTeX

@inproceedings{tucker2017iclr-rebar,
  title     = {{REBAR: Low-Variance, Unbiased Gradient Estimates for Discrete Latent Variable Models}},
  author    = {Tucker, George and Mnih, Andriy and Maddison, Chris J. and Sohl-Dickstein, Jascha},
  booktitle = {International Conference on Learning Representations},
  year      = {2017},
  url       = {https://mlanthology.org/iclr/2017/tucker2017iclr-rebar/}
}