Quantitative Convergence of Trained Neural Networks to Gaussian Processes

Abstract

In this paper, we study the quantitative convergence of shallow neural networks trained via gradient descent to their associated Gaussian processes in the infinite-width limit. While previous work has established qualitative convergence under broad settings, precise, finite-width estimates remain limited, particularly during training. We provide explicit upper bounds on the quadratic Wasserstein distance between the network output and its Gaussian approximation at any training time $t \ge 0$, demonstrating polynomial decay with network width. Our results quantify how architectural parameters, such as width and input dimension, influence convergence, and how training dynamics affect the approximation error

Cite

Text

Agazzi et al. "Quantitative Convergence of Trained Neural Networks to Gaussian Processes." Advances in Neural Information Processing Systems, 2025.

Markdown

[Agazzi et al. "Quantitative Convergence of Trained Neural Networks to Gaussian Processes." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/agazzi2025neurips-quantitative/)

BibTeX

@inproceedings{agazzi2025neurips-quantitative,
  title     = {{Quantitative Convergence of Trained Neural Networks to Gaussian Processes}},
  author    = {Agazzi, Andrea and García, Eloy Mosig and Trevisan, Dario},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/agazzi2025neurips-quantitative/}
}