Boosting in the Limit: Maximizing the Margin of Learned Ensembles

Abstract

The "minimum margin" of an ensemble classifier on a given training set is, roughly speaking, the smallest vote it gives to any correct training label. Recent work has shown that the Adaboost algorithm is particularly effective at producing ensembles with large minimum margins, and theory suggests that this may account for its success at reducing generalization error. We note, however, that the problem of finding good margins is closely related to linear programming, and we use this connection to derive and test new "LPboosting" algorithms that achieve better minimum margins than Adaboost. However, these algorithms do not always yield better generalization performance. In fact, more often the opposite is true. We report on a series of controlled experiments which show that no simple version of the minimum-margin story can be complete. We conclude that the crucial question as to why boosting works so well in practice, and how to further improve upon it, remains mostly open. Some of our ...

Cite

Text

Grove and Schuurmans. "Boosting in the Limit: Maximizing the Margin of Learned Ensembles." AAAI Conference on Artificial Intelligence, 1998.

Markdown

[Grove and Schuurmans. "Boosting in the Limit: Maximizing the Margin of Learned Ensembles." AAAI Conference on Artificial Intelligence, 1998.](https://mlanthology.org/aaai/1998/grove1998aaai-boosting/)

BibTeX

@inproceedings{grove1998aaai-boosting,
  title     = {{Boosting in the Limit: Maximizing the Margin of Learned Ensembles}},
  author    = {Grove, Adam J. and Schuurmans, Dale},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {1998},
  pages     = {692-699},
  url       = {https://mlanthology.org/aaai/1998/grove1998aaai-boosting/}
}