The Transition to Perfect Generalization in Perceptrons

Abstract

Several recent papers (Gardner and Derrida 1989; Györgyi 1990; Sompolinsky et al. 1990) have found, using methods of statistical physics, that a transition to perfect generalization occurs in training a simple perceptron whose weights can only take values ±1. We give a rigorous proof of such a phenomena. That is, we show, for α = 2.0821, that if at least αn examples are drawn from the uniform distribution on +1, −1n and classified according to a target perceptron wt ∈ +1, −1n as positive or negative according to whether wt·x is nonnegative or negative, then the probability is 2−(√n) that there is any other such perceptron consistent with the examples. Numerical results indicate further that perfect generalization holds for α as low as 1.5.

Cite

Text

Baum and Lyuu. "The Transition to Perfect Generalization in Perceptrons." Neural Computation, 1991. doi:10.1162/NECO.1991.3.3.386

Markdown

[Baum and Lyuu. "The Transition to Perfect Generalization in Perceptrons." Neural Computation, 1991.](https://mlanthology.org/neco/1991/baum1991neco-transition/) doi:10.1162/NECO.1991.3.3.386

BibTeX

@article{baum1991neco-transition,
  title     = {{The Transition to Perfect Generalization in Perceptrons}},
  author    = {Baum, Eric B. and Lyuu, Yuh-Dauh},
  journal   = {Neural Computation},
  year      = {1991},
  pages     = {386-401},
  doi       = {10.1162/NECO.1991.3.3.386},
  volume    = {3},
  url       = {https://mlanthology.org/neco/1991/baum1991neco-transition/}
}