AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC
Abstract
Stochastic gradient Hamiltonian Monte Carlo (SGHMC) is an efficient method for sampling from continuous distributions. It is a faster alternative to HMC: instead of using the whole dataset at each iteration, SGHMC uses only a subsample. This improves performance, but introduces bias that can cause SGHMC to converge to the wrong distribution. One can prevent this using a step size that decays to zero, but such a step size schedule can drastically slow down convergence. To address this tension, we propose a novel second-order SG-MCMC algorithm—AMAGOLD—that infrequently uses Metropolis-Hastings (M-H) corrections to remove bias. The infrequency of corrections amortizes their cost. We prove AMAGOLD converges to the target distribution with a fixed, rather than a diminishing, step size, and that its convergence rate is at most a constant factor slower than a full-batch baseline. We empirically demonstrate AMAGOLD’s effectiveness on synthetic distributions, Bayesian logistic regression, and Bayesian neural networks.
Cite
Text
Zhang et al. "AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC." Artificial Intelligence and Statistics, 2020.Markdown
[Zhang et al. "AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC." Artificial Intelligence and Statistics, 2020.](https://mlanthology.org/aistats/2020/zhang2020aistats-amagold/)BibTeX
@inproceedings{zhang2020aistats-amagold,
title = {{AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC}},
author = {Zhang, Ruqi and Cooper, A. Feder and De Sa, Christopher},
booktitle = {Artificial Intelligence and Statistics},
year = {2020},
pages = {2142-2152},
volume = {108},
url = {https://mlanthology.org/aistats/2020/zhang2020aistats-amagold/}
}