MCMC Should Mix: Learning Energy-Based Model with Neural Transport Latent Space MCMC
Abstract
Learning energy-based model (EBM) requires MCMC sampling of the learned model as an inner loop of the learning algorithm. However, MCMC sampling of EBMs in high-dimensional data space is generally not mixing, because the energy function, which is usually parametrized by deep network, is highly multi-modal in the data space. This is a serious handicap for both theory and practice of EBMs. In this paper, we propose to learn EBM with a flow-based model (or in general latent variable model) serving as a backbone, so that the EBM is a correction or an exponential tilting of the flow-based model. We show that the model has a particularly simple form in the space of the latent variables of the generative model, and MCMC sampling of the EBM in the latent space mixes well and traverses modes in the data space. This enables proper sampling and learning of EBMs.
Cite
Text
Nijkamp et al. "MCMC Should Mix: Learning Energy-Based Model with Neural Transport Latent Space MCMC." International Conference on Learning Representations, 2022.Markdown
[Nijkamp et al. "MCMC Should Mix: Learning Energy-Based Model with Neural Transport Latent Space MCMC." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/nijkamp2022iclr-mcmc/)BibTeX
@inproceedings{nijkamp2022iclr-mcmc,
title = {{MCMC Should Mix: Learning Energy-Based Model with Neural Transport Latent Space MCMC}},
author = {Nijkamp, Erik and Gao, Ruiqi and Sountsov, Pavel and Vasudevan, Srinivas and Pang, Bo and Zhu, Song-Chun and Wu, Ying Nian},
booktitle = {International Conference on Learning Representations},
year = {2022},
url = {https://mlanthology.org/iclr/2022/nijkamp2022iclr-mcmc/}
}