ELBO, Regularized Maximum Likelihood, and Their Common One-Sample Approximation for Training Stochastic Neural Networks

Abstract

Monte Carlo approximations are central to the training of stochastic neural networks in general, and Bayesian neural networks (BNNs) in particular. We observe that the common one-sample approximation of the standard training objective can be viewed both as maximizing the Evidence Lower Bound (ELBO) and as maximizing a regularized log-likelihood of a compound distribution. This latter approach differs from the ELBO only in the order of the logarithm and expectation, and is theoretically grounded in PAC-Bayes theory. We argue theoretically and demonstrate empirically that training with the regularized maximum likelihood increases prediction variance, enhancing performance in misspecified settings, adversarial robustness, and strengthening out-of-distribution (OOD) detection. Our findings help reconcile previous contradictions in the literature by providing a detailed analysis of how training objectives and Monte Carlo sample sizes affect uncertainty quantification in stochastic neural networks.

Cite

Text

Däubener et al. "ELBO, Regularized Maximum Likelihood, and Their Common One-Sample Approximation for Training Stochastic Neural Networks." Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence, 2025.

Markdown

[Däubener et al. "ELBO, Regularized Maximum Likelihood, and Their Common One-Sample Approximation for Training Stochastic Neural Networks." Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence, 2025.](https://mlanthology.org/uai/2025/daubener2025uai-elbo/)

BibTeX

@inproceedings{daubener2025uai-elbo,
  title     = {{ELBO, Regularized Maximum Likelihood, and Their Common One-Sample Approximation for Training Stochastic Neural Networks}},
  author    = {Däubener, Sina and Damm, Simon and Fischer, Asja},
  booktitle = {Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence},
  year      = {2025},
  pages     = {897-914},
  volume    = {286},
  url       = {https://mlanthology.org/uai/2025/daubener2025uai-elbo/}
}