A Quantitative Study of Small Disjuncts

Abstract

Systems that learn from examples often express the learned concept in the form of a disjunctive description. Disjuncts that correctly classify few training examples are known as small disjuncts and are interesting to machine learning researchers because they have a much higher error rate than large disjuncts. Previous research has investigated this phenomenon by performing ad hoc analyses of a small number of datasets. In this paper we present a quantitative measure for evaluating the effect of small disjuncts on learning and use it to analyze 30 benchmark datasets. We investigate the relationship between small disjuncts and pruning, training set size and noise, and come up with several interesting results. Introduction Systems that learn from examples often express the learned concept as a disjunction. The size of a disjunct is defined as the number of training examples that it correctly classifies (Holte, Acker, and Porter 1989). A number of empirical studies have demons...

Cite

Text

Weiss and Hirsh. "A Quantitative Study of Small Disjuncts." AAAI Conference on Artificial Intelligence, 2000.

Markdown

[Weiss and Hirsh. "A Quantitative Study of Small Disjuncts." AAAI Conference on Artificial Intelligence, 2000.](https://mlanthology.org/aaai/2000/weiss2000aaai-quantitative/)

BibTeX

@inproceedings{weiss2000aaai-quantitative,
  title     = {{A Quantitative Study of Small Disjuncts}},
  author    = {Weiss, Gary M. and Hirsh, Haym},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2000},
  pages     = {665-670},
  url       = {https://mlanthology.org/aaai/2000/weiss2000aaai-quantitative/}
}