On the Natural Gradient of the Evidence Lower Bound

Abstract

This article studies the Fisher-Rao gradient, also referred to as the natural gradient, of the evidence lower bound (ELBO) which plays a central role in generative machine learning. It reveals that the gap between the evidence and its lower bound, the ELBO, has essentially a vanishing natural gradient within unconstrained optimization. As a result, maximization of the ELBO is equivalent to minimization of the Kullback-Leibler divergence from a target distribution, the primary objective function of learning. Building on this insight, we derive a condition under which this equivalence persists even when optimization is constrained to a model. This condition yields a geometric characterization, which we formalize through the notion of a cylindrical model.

Cite

Text

Ay et al. "On the Natural Gradient of the Evidence Lower Bound." Journal of Machine Learning Research, 2025.

Markdown

[Ay et al. "On the Natural Gradient of the Evidence Lower Bound." Journal of Machine Learning Research, 2025.](https://mlanthology.org/jmlr/2025/ay2025jmlr-natural/)

BibTeX

@article{ay2025jmlr-natural,
  title     = {{On the Natural Gradient of the Evidence Lower Bound}},
  author    = {Ay, Nihat and van Oostrum, Jesse and Datar, Adwait},
  journal   = {Journal of Machine Learning Research},
  year      = {2025},
  pages     = {1-37},
  volume    = {26},
  url       = {https://mlanthology.org/jmlr/2025/ay2025jmlr-natural/}
}