Enhanced Gradient-Based MCMC in Discrete Spaces
Abstract
The recent introduction of gradient-based Markov chain Monte Carlo (MCMC) for discrete spaces holds great promise, and comes with the tantalising possibility of new discrete counterparts to celebrated continuous methods such as the Metropolis-adjusted Langevin algorithm (MALA). Towards this goal, we introduce several discrete Metropolis-Hastings samplers that are conceptually inspired by MALA, and demonstrate their strong empirical performance across a range of challenging sampling problems in Bayesian inference and energy-based modelling. Methodologically, we identify why discrete analogues to \emph{preconditioned} MALA are generally intractable, motivating us to introduce a new kind of preconditioning based on auxiliary variables and the `Gaussian integral trick'.
Cite
Text
Rhodes and Gutmann. "Enhanced Gradient-Based MCMC in Discrete Spaces." Transactions on Machine Learning Research, 2022.Markdown
[Rhodes and Gutmann. "Enhanced Gradient-Based MCMC in Discrete Spaces." Transactions on Machine Learning Research, 2022.](https://mlanthology.org/tmlr/2022/rhodes2022tmlr-enhanced/)BibTeX
@article{rhodes2022tmlr-enhanced,
title = {{Enhanced Gradient-Based MCMC in Discrete Spaces}},
author = {Rhodes, Benjamin and Gutmann, Michael U.},
journal = {Transactions on Machine Learning Research},
year = {2022},
url = {https://mlanthology.org/tmlr/2022/rhodes2022tmlr-enhanced/}
}