Meta-Learning MCMC Proposals

Abstract

Effective implementations of sampling-based probabilistic inference often require manually constructed, model-specific proposals. Inspired by recent progresses in meta-learning for training learning agents that can generalize to unseen environments, we propose a meta-learning approach to building effective and generalizable MCMC proposals. We parametrize the proposal as a neural network to provide fast approximations to block Gibbs conditionals. The learned neural proposals generalize to occurrences of common structural motifs across different models, allowing for the construction of a library of learned inference primitives that can accelerate inference on unseen models with no model-specific training required. We explore several applications including open-universe Gaussian mixture models, in which our learned proposals outperform a hand-tuned sampler, and a real-world named entity recognition task, in which our sampler yields higher final F1 scores than classical single-site Gibbs sampling.

Cite

Text

Wang et al. "Meta-Learning MCMC Proposals." Neural Information Processing Systems, 2018.

Markdown

[Wang et al. "Meta-Learning MCMC Proposals." Neural Information Processing Systems, 2018.](https://mlanthology.org/neurips/2018/wang2018neurips-metalearning/)

BibTeX

@inproceedings{wang2018neurips-metalearning,
  title     = {{Meta-Learning MCMC Proposals}},
  author    = {Wang, Tongzhou and Wu, Yi and Moore, Dave and Russell, Stuart},
  booktitle = {Neural Information Processing Systems},
  year      = {2018},
  pages     = {4146-4156},
  url       = {https://mlanthology.org/neurips/2018/wang2018neurips-metalearning/}
}