On Contrastive Learning for Likelihood-Free Inference
Abstract
Likelihood-free methods perform parameter inference in stochastic simulator models where evaluating the likelihood is intractable but sampling synthetic data is possible. One class of methods for this likelihood-free problem uses a classifier to distinguish between pairs of parameter-observation samples generated using the simulator and pairs sampled from some reference distribution, which implicitly learns a density ratio proportional to the likelihood. Another popular class of methods fits a conditional distribution to the parameter posterior directly, and a particular recent variant allows for the use of flexible neural density estimators for this task. In this work, we show that both of these approaches can be unified under a general contrastive learning scheme, and clarify how they should be run and compared.
Cite
Text
Durkan et al. "On Contrastive Learning for Likelihood-Free Inference." International Conference on Machine Learning, 2020.Markdown
[Durkan et al. "On Contrastive Learning for Likelihood-Free Inference." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/durkan2020icml-contrastive/)BibTeX
@inproceedings{durkan2020icml-contrastive,
title = {{On Contrastive Learning for Likelihood-Free Inference}},
author = {Durkan, Conor and Murray, Iain and Papamakarios, George},
booktitle = {International Conference on Machine Learning},
year = {2020},
pages = {2771-2781},
volume = {119},
url = {https://mlanthology.org/icml/2020/durkan2020icml-contrastive/}
}