Random Fully Connected Neural Networks as Perturbatively Solvable Hierarchies

Abstract

We study the distribution of fully connected neural networks with Gaussian random weights/biases and L hidden layers, each of width proportional to a large parameter n. For polynomially bounded non-linearities we give sharp estimates in powers of 1/n for the joint cumulants of the network output and its derivatives. We further show that network cumulants form a perturbatively solvable hierarchy in powers of 1/n. That is, the k-th order cumulants in each layer are determined to leading order in 1/n by cumulants of order at most k computed at the previous layer. By explicitly deriving and then solving several such recursions, we find that the depth-to-width ratio L/n plays the role of an effective network depth, controlling both the distance to Gaussianity and the size of inter-neuron correlations.

Cite

Text

Hanin. "Random Fully Connected Neural Networks as Perturbatively Solvable Hierarchies." Journal of Machine Learning Research, 2024.

Markdown

[Hanin. "Random Fully Connected Neural Networks as Perturbatively Solvable Hierarchies." Journal of Machine Learning Research, 2024.](https://mlanthology.org/jmlr/2024/hanin2024jmlr-random/)

BibTeX

@article{hanin2024jmlr-random,
  title     = {{Random Fully Connected Neural Networks as Perturbatively Solvable Hierarchies}},
  author    = {Hanin, Boris},
  journal   = {Journal of Machine Learning Research},
  year      = {2024},
  pages     = {1-58},
  volume    = {25},
  url       = {https://mlanthology.org/jmlr/2024/hanin2024jmlr-random/}
}