Rigorous Learning Curve Bounds from Statistical Mechanics
Abstract
In this paper we introduce and investigate a mathematically rigorous theory of learning curves that is based on ideas from statistical mechanics. The advantage of our theory over the well-established Vapnik-Chervonenkis theory is that our bounds can be considerably tighter in many cases, and are also more reflective of the true behavior of learning curves. This behavior can often exhibit dramatic properties such as phase transitions, as well as power law asymptotics not explained by the VC theory. The disadvantages of our theory are that its application requires knowledge of the input distribution, and it is limited so far to finite cardinality function classes. We illustrate our results with many concrete examples of learning curve bounds derived from our theory.
Cite
Text
Haussler et al. "Rigorous Learning Curve Bounds from Statistical Mechanics." Machine Learning, 1996. doi:10.1023/A:1026499208981Markdown
[Haussler et al. "Rigorous Learning Curve Bounds from Statistical Mechanics." Machine Learning, 1996.](https://mlanthology.org/mlj/1996/haussler1996mlj-rigorous/) doi:10.1023/A:1026499208981BibTeX
@article{haussler1996mlj-rigorous,
title = {{Rigorous Learning Curve Bounds from Statistical Mechanics}},
author = {Haussler, David and Kearns, Michael J. and Seung, H. Sebastian and Tishby, Naftali},
journal = {Machine Learning},
year = {1996},
pages = {195-236},
doi = {10.1023/A:1026499208981},
volume = {25},
url = {https://mlanthology.org/mlj/1996/haussler1996mlj-rigorous/}
}