Meta-Learning MCMC Proposals
Abstract
Effective implementations of sampling-based probabilistic inference often require manually constructed, model-specific proposals. Inspired by recent progresses in meta-learning for training learning agents that can generalize to unseen environments, we propose a meta-learning approach to building effective and generalizable MCMC proposals. We parametrize the proposal as a neural network to provide fast approximations to block Gibbs conditionals. The learned neural proposals generalize to occurrences of common structural motifs across different models, allowing for the construction of a library of learned inference primitives that can accelerate inference on unseen models with no model-specific training required. We explore several applications including open-universe Gaussian mixture models, in which our learned proposals outperform a hand-tuned sampler, and a real-world named entity recognition task, in which our sampler yields higher final F1 scores than classical single-site Gibbs sampling.
Cite
Text
Wang et al. "Meta-Learning MCMC Proposals." Neural Information Processing Systems, 2018.Markdown
[Wang et al. "Meta-Learning MCMC Proposals." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/wang2018neurips-metalearning/)BibTeX
@inproceedings{wang2018neurips-metalearning,
title = {{Meta-Learning MCMC Proposals}},
author = {Wang, Tongzhou and Wu, Yi and Moore, Dave and Russell, Stuart},
booktitle = {Neural Information Processing Systems},
year = {2018},
pages = {4146-4156},
url = {https://mlanthology.org/neurips/2018/wang2018neurips-metalearning/}
}