Large Language Bayes
Abstract
Many domain experts do not have the time or expertise to write formal Bayesian models. This paper takes an informal problem description as input, and combines a large language model and a probabilistic programming language to define a joint distribution over formal models, latent variables, and data. A posterior over latent variables follows by conditioning on observed data and integrating over formal models. This presents a challenging inference problem. We suggest an inference recipe that amounts to generating many formal models from the large language model, performing approximate inference on each, and then doing a weighted aver- age. This is justified and analyzed as a combination of self-normalized importance sampling, MCMC, and importance-weighted variational inference. Experimentally, this produces sensible predictions from only data and an informal problem description, without the need to specify a formal model.
Cite
Text
Domke. "Large Language Bayes." Advances in Neural Information Processing Systems, 2025.Markdown
[Domke. "Large Language Bayes." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/domke2025neurips-large/)BibTeX
@inproceedings{domke2025neurips-large,
title = {{Large Language Bayes}},
author = {Domke, Justin},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/domke2025neurips-large/}
}