Online Learning from Finite Training Sets: An Analytical Case Study
Abstract
We analyse online learning from finite training sets at non(cid:173) infinitesimal learning rates TJ. By an extension of statistical me(cid:173) chanics methods, we obtain exact results for the time-dependent generalization error of a linear network with a large number of weights N. We find, for example, that for small training sets of size p ~ N, larger learning rates can be used without compromis(cid:173) ing asymptotic generalization performance or convergence speed. Encouragingly, for optimal settings of TJ (and, less importantly, weight decay ,) at given final learning time, the generalization per(cid:173) formance of online learning is essentially as good as that of offline learning.
Cite
Text
Sollich and Barber. "Online Learning from Finite Training Sets: An Analytical Case Study." Neural Information Processing Systems, 1996.Markdown
[Sollich and Barber. "Online Learning from Finite Training Sets: An Analytical Case Study." Neural Information Processing Systems, 1996.](https://mlanthology.org/neurips/1996/sollich1996neurips-online/)BibTeX
@inproceedings{sollich1996neurips-online,
title = {{Online Learning from Finite Training Sets: An Analytical Case Study}},
author = {Sollich, Peter and Barber, David},
booktitle = {Neural Information Processing Systems},
year = {1996},
pages = {274-280},
url = {https://mlanthology.org/neurips/1996/sollich1996neurips-online/}
}