An Algorithmic Theory of Learning: Robust Concepts and Random Projection

Abstract

We study the phenomenon of cognitive learning from an algorithmic standpoint. How does the brain effectively learn concepts from a small number of examples despite the fact that each example contains a huge amount of information? We provide a novel algorithmic analysis via a model of robust concept learning (closely related to “margin classifiers”), and show that a relatively small number of examples are sufficient to learn rich concept classes. The new algorithms have several advantages—they are faster, conceptually simpler, and resistant to low levels of noise. For example, a robust half-space can be learned in linear time using only a constant number of training examples, regardless of the number of attributes. A general (algorithmic) consequence of the model, that “more robust concepts are easier to learn”, is supported by a multitude of psychological studies.

Cite

Text

Arriaga and Vempala. "An Algorithmic Theory of Learning: Robust Concepts and Random Projection." Machine Learning, 2006. doi:10.1007/S10994-006-6265-7

Markdown

[Arriaga and Vempala. "An Algorithmic Theory of Learning: Robust Concepts and Random Projection." Machine Learning, 2006.](https://mlanthology.org/mlj/2006/arriaga2006mlj-algorithmic/) doi:10.1007/S10994-006-6265-7

BibTeX

@article{arriaga2006mlj-algorithmic,
  title     = {{An Algorithmic Theory of Learning: Robust Concepts and Random Projection}},
  author    = {Arriaga, Rosa I. and Vempala, Santosh S.},
  journal   = {Machine Learning},
  year      = {2006},
  pages     = {161-182},
  doi       = {10.1007/S10994-006-6265-7},
  volume    = {63},
  url       = {https://mlanthology.org/mlj/2006/arriaga2006mlj-algorithmic/}
}