Average-Case Analysis of a Nearest Neighbor Algorithm

Abstract

In this paper we present an average-case analysis of the nearest neighbor algorithm, a simple induction method that has been studied by many researchers. Our analysis assumes a conjunctive target concept, noise-free Boolean attributes, and a uniform distribution over the instance space. We calculate the probability that the algorithm will encounter a test instance that is distance d from the prototype of the concept, along with the probability that the nearest stored training case is distance e from this test instance. From this we compute the probability of correct classification as a function of the number of observed training cases, the number of relevant attributes, and the number of irrelevant attributes. We also explore the behavioral implications of the analysis by presenting predicted learning curves for artificial domains, and give experimental results on these domains as a check on our reasoning.

Cite

Text

Langley and Iba. "Average-Case Analysis of a Nearest Neighbor Algorithm." International Joint Conference on Artificial Intelligence, 1993.

Markdown

[Langley and Iba. "Average-Case Analysis of a Nearest Neighbor Algorithm." International Joint Conference on Artificial Intelligence, 1993.](https://mlanthology.org/ijcai/1993/langley1993ijcai-average/)

BibTeX

@inproceedings{langley1993ijcai-average,
  title     = {{Average-Case Analysis of a Nearest Neighbor Algorithm}},
  author    = {Langley, Pat and Iba, Wayne},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {1993},
  pages     = {889-894},
  url       = {https://mlanthology.org/ijcai/1993/langley1993ijcai-average/}
}