Temporal Dynamics of Generalization in Neural Networks

Abstract

This paper presents a rigorous characterization of how a general nonlinear learning machine generalizes during the training process when it is trained on a random sample using a gradient descent algorithm based on reduction of training error. It is shown, in particular, that best generalization performance occurs, in general, before the global minimum of the training error is achieved. The different roles played by the complexity of the machine class and the complexity of the specific machine in the class during learning are also precisely demarcated.

Cite

Text

Wang and Venkatesh. "Temporal Dynamics of Generalization in Neural Networks." Neural Information Processing Systems, 1994.

Markdown

[Wang and Venkatesh. "Temporal Dynamics of Generalization in Neural Networks." Neural Information Processing Systems, 1994.](https://mlanthology.org/neurips/1994/wang1994neurips-temporal/)

BibTeX

@inproceedings{wang1994neurips-temporal,
  title     = {{Temporal Dynamics of Generalization in Neural Networks}},
  author    = {Wang, Changfeng and Venkatesh, Santosh S.},
  booktitle = {Neural Information Processing Systems},
  year      = {1994},
  pages     = {263-270},
  url       = {https://mlanthology.org/neurips/1994/wang1994neurips-temporal/}
}