Policy Gradients for Optimal Parallel Tempering MCMC
Abstract
Parallel tempering is a meta-algorithm for Markov Chain Monte Carlo (MCMC) methods which uses multiple chains to sample from tempered versions of the target distribution, improving mixing on multi-modal distributions that are difficult to explore for traditional methods. The success of this technique depends critically on the choice of chain temperatures. We introduce an adaptive temperature selection algorithm which adjusts temperatures during sampling using a policy gradient method. Experimental results show that it can outperform traditional geometrically-spaced temperatures and uniform acceptance rate temperature ladders in terms of integrated autocorrelation time on test distributions.
Cite
Text
Zhao and Pillai. "Policy Gradients for Optimal Parallel Tempering MCMC." ICML 2024 Workshops: SPIGM, 2024.Markdown
[Zhao and Pillai. "Policy Gradients for Optimal Parallel Tempering MCMC." ICML 2024 Workshops: SPIGM, 2024.](https://mlanthology.org/icmlw/2024/zhao2024icmlw-policy/)BibTeX
@inproceedings{zhao2024icmlw-policy,
title = {{Policy Gradients for Optimal Parallel Tempering MCMC}},
author = {Zhao, Daniel and Pillai, Natesh S.},
booktitle = {ICML 2024 Workshops: SPIGM},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/zhao2024icmlw-policy/}
}