Stochastic Gradient Monomial Gamma Sampler

Abstract

Scaling Markov Chain Monte Carlo (MCMC) to estimate posterior distributions from large datasets has been made possible as a result of advances in stochastic gradient techniques. Despite their success, mixing performance of existing methods when sampling from multimodal distributions can be less efficient with insufficient Monte Carlo samples; this is evidenced by slow convergence and insufficient exploration of posterior distributions. We propose a generalized framework to improve the sampling efficiency of stochastic gradient MCMC, by leveraging a generalized kinetics that delivers superior stationary mixing, especially in multimodal distributions, and propose several techniques to overcome the practical issues. We show that the proposed approach is better at exploring a complicated multimodal posterior distribution, and demonstrate improvements over other stochastic gradient MCMC methods on various applications.

Cite

Text

Zhang et al. "Stochastic Gradient Monomial Gamma Sampler." International Conference on Machine Learning, 2017.

Markdown

[Zhang et al. "Stochastic Gradient Monomial Gamma Sampler." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/zhang2017icml-stochastic/)

BibTeX

@inproceedings{zhang2017icml-stochastic,
  title     = {{Stochastic Gradient Monomial Gamma Sampler}},
  author    = {Zhang, Yizhe and Chen, Changyou and Gan, Zhe and Henao, Ricardo and Carin, Lawrence},
  booktitle = {International Conference on Machine Learning},
  year      = {2017},
  pages     = {3996-4005},
  volume    = {70},
  url       = {https://mlanthology.org/icml/2017/zhang2017icml-stochastic/}
}