Learning to Explore for Stochastic Gradient MCMC
Abstract
Bayesian Neural Networks(BNNs) with high-dimensional parameters pose a challenge for posterior inference due to the multi-modality of the posterior distributions. Stochastic Gradient Markov Chain Monte Carlo(SGMCMC) with cyclical learning rate scheduling is a promising solution, but it requires a large number of sampling steps to explore high-dimensional multi-modal posteriors, making it computationally expensive. In this paper, we propose a meta-learning strategy to build SGMCMC which can efficiently explore the multi-modal target distributions. Our algorithm allows the learned SGMCMC to quickly explore the high-density region of the posterior landscape. Also, we show that this exploration property is transferrable to various tasks, even for the ones unseen during a meta-training stage. Using popular image classification benchmarks and a variety of downstream tasks, we demonstrate that our method significantly improves the sampling efficiency, achieving better performance than vanilla SGMCMC without incurring significant computational overhead.
Cite
Text
Kim et al. "Learning to Explore for Stochastic Gradient MCMC." International Conference on Machine Learning, 2024.Markdown
[Kim et al. "Learning to Explore for Stochastic Gradient MCMC." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/kim2024icml-learning/)BibTeX
@inproceedings{kim2024icml-learning,
title = {{Learning to Explore for Stochastic Gradient MCMC}},
author = {Kim, Seunghyun and Jung, Seohyeon and Kim, Seonghyeon and Lee, Juho},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {24015-24039},
volume = {235},
url = {https://mlanthology.org/icml/2024/kim2024icml-learning/}
}