Oops I Took a Gradient: Scalable Sampling for Discrete Distributions

Abstract

We propose a general and scalable approximate sampling strategy for probabilistic models with discrete variables. Our approach uses gradients of the likelihood function with respect to its discrete inputs to propose updates in a Metropolis-Hastings sampler. We show empirically that this approach outperforms generic samplers in a number of difficult settings including Ising models, Potts models, restricted Boltzmann machines, and factorial hidden Markov models. We also demonstrate our improved sampler for training deep energy-based models on high dimensional discrete image data. This approach outperforms variational auto-encoders and existing energy-based models. Finally, we give bounds showing that our approach is near-optimal in the class of samplers which propose local updates.

Cite

Text

Grathwohl et al. "Oops I Took a Gradient: Scalable Sampling for Discrete Distributions." International Conference on Machine Learning, 2021.

Markdown

[Grathwohl et al. "Oops I Took a Gradient: Scalable Sampling for Discrete Distributions." International Conference on Machine Learning, 2021.](https://mlanthology.org/icml/2021/grathwohl2021icml-oops/)

BibTeX

@inproceedings{grathwohl2021icml-oops,
  title     = {{Oops I Took a Gradient: Scalable Sampling for Discrete Distributions}},
  author    = {Grathwohl, Will and Swersky, Kevin and Hashemi, Milad and Duvenaud, David and Maddison, Chris},
  booktitle = {International Conference on Machine Learning},
  year      = {2021},
  pages     = {3831-3841},
  volume    = {139},
  url       = {https://mlanthology.org/icml/2021/grathwohl2021icml-oops/}
}