A Theory for Memory-Based Learning
Abstract
A memory-based learning system is an extended memory management system that decomposes the input space either statically or dynamically into subregions for the purpose of storing and retrieving functional information. The main generalization techniques employed by memory-based learning systems are the nearest-neighbor search, space decomposition techniques, and clustering. Research on memory-based learning is still in its early stage. In particular, there are very few rigorous theoretical results regarding memory requirement, sample size, expected performance, and computational complexity. In this paper, we propose a model for memory-based learning and use it to analyze several methods— ε-covering, hashing, clustering, tree-structured clustering, and receptive-fields— for learning smooth functions. The sample size and system complexity are derived for each method. Our model is built upon the generalized PAC learning model of Haussler and is closely related to the method of vector quantization in data compression. Our main result is that we can build memory-based learning systems using new clustering algorithms [LiVb] to PAC-learn in polynomial time using only polynomial storage in typical situations.
Cite
Text
Lin and Vitter. "A Theory for Memory-Based Learning." Annual Conference on Computational Learning Theory, 1992. doi:10.1145/130385.130397Markdown
[Lin and Vitter. "A Theory for Memory-Based Learning." Annual Conference on Computational Learning Theory, 1992.](https://mlanthology.org/colt/1992/lin1992colt-theory/) doi:10.1145/130385.130397BibTeX
@inproceedings{lin1992colt-theory,
title = {{A Theory for Memory-Based Learning}},
author = {Lin, Jyh-Han and Vitter, Jeffrey Scott},
booktitle = {Annual Conference on Computational Learning Theory},
year = {1992},
pages = {103-115},
doi = {10.1145/130385.130397},
url = {https://mlanthology.org/colt/1992/lin1992colt-theory/}
}