Online Learning from Finite Training Sets: An Analytical Case Study

Abstract

We analyse online learning from finite training sets at non(cid:173) infinitesimal learning rates TJ. By an extension of statistical me(cid:173) chanics methods, we obtain exact results for the time-dependent generalization error of a linear network with a large number of weights N. We find, for example, that for small training sets of size p ~ N, larger learning rates can be used without compromis(cid:173) ing asymptotic generalization performance or convergence speed. Encouragingly, for optimal settings of TJ (and, less importantly, weight decay ,) at given final learning time, the generalization per(cid:173) formance of online learning is essentially as good as that of offline learning.

Cite

Text

Sollich and Barber. "Online Learning from Finite Training Sets: An Analytical Case Study." Neural Information Processing Systems, 1996.

Markdown

[Sollich and Barber. "Online Learning from Finite Training Sets: An Analytical Case Study." Neural Information Processing Systems, 1996.](https://mlanthology.org/neurips/1996/sollich1996neurips-online/)

BibTeX

@inproceedings{sollich1996neurips-online,
  title     = {{Online Learning from Finite Training Sets: An Analytical Case Study}},
  author    = {Sollich, Peter and Barber, David},
  booktitle = {Neural Information Processing Systems},
  year      = {1996},
  pages     = {274-280},
  url       = {https://mlanthology.org/neurips/1996/sollich1996neurips-online/}
}