Boosting the Margin: A New Explanation for the Effectiveness of Voting Methods

Abstract

One of the surprising recurring phenomena observed in experiments\nwith boosting is that the test error of the generated classifier usually does\nnot increase as its size becomes very large, and often is observed to decrease\neven after the training error reaches zero. In this paper, we show that this\nphenomenon is related to the distribution of margins of the training\nexamples with respect to the generated voting classification rule, where the\nmargin of an example is simply the difference between the number of correct\nvotes and the maximum number of votes received by any incorrect label. We show\nthat techniques used in the analysis of Vapnik’s support vector\nclassifiers and of neural networks with small weights can be applied to voting\nmethods to relate the margin distribution to the test error. We also show\ntheoretically and experimentally that boosting is especially effective at\nincreasing the margins of the training examples. Finally, we compare our\nexplanation to those based on the bias-variance decomposition.

Cite

Text

Schapire et al. "Boosting the Margin: A New Explanation for the Effectiveness of Voting Methods." International Conference on Machine Learning, 1997. doi:10.1214/AOS/1024691352

Markdown

[Schapire et al. "Boosting the Margin: A New Explanation for the Effectiveness of Voting Methods." International Conference on Machine Learning, 1997.](https://mlanthology.org/icml/1997/schapire1997icml-boosting/) doi:10.1214/AOS/1024691352

BibTeX

@inproceedings{schapire1997icml-boosting,
  title     = {{Boosting the Margin: A New Explanation for the Effectiveness of Voting Methods}},
  author    = {Schapire, Robert E. and Freund, Yoav and Barlett, Peter and Lee, Wee Sun},
  booktitle = {International Conference on Machine Learning},
  year      = {1997},
  pages     = {322-330},
  doi       = {10.1214/AOS/1024691352},
  url       = {https://mlanthology.org/icml/1997/schapire1997icml-boosting/}
}