Totally Corrective Boosting Algorithms That Maximize the Margin
Abstract
We consider boosting algorithms that maintain a distribution over a set of examples. At each iteration a weak hypothesis is received and the distribution is updated. We motivate these updates as minimizing the relative entropy subject to linear constraints. For example AdaBoost constrains the edge of the last hypothesis w.r.t. the updated distribution to be at most γ = 0. In some sense, AdaBoost is "corrective" w.r.t. the last hypothesis. A cleaner boosting method is to be "totally corrective": the edges of all past hypotheses are constrained to be at most γ, where γ is suitably adapted. Using new techniques, we prove the same iteration bounds for the totally corrective algorithms as for their corrective versions. Moreover with adaptive γ, the algorithms provably maximizes the margin. Experimentally, the totally corrective versions return smaller convex combinations of weak hypotheses than the corrective ones and are competitive with LPBoost, a totally corrective boosting algorithm with no regularization, for which there is no iteration bound known.
Cite
Text
Warmuth et al. "Totally Corrective Boosting Algorithms That Maximize the Margin." International Conference on Machine Learning, 2006. doi:10.1145/1143844.1143970Markdown
[Warmuth et al. "Totally Corrective Boosting Algorithms That Maximize the Margin." International Conference on Machine Learning, 2006.](https://mlanthology.org/icml/2006/warmuth2006icml-totally/) doi:10.1145/1143844.1143970BibTeX
@inproceedings{warmuth2006icml-totally,
title = {{Totally Corrective Boosting Algorithms That Maximize the Margin}},
author = {Warmuth, Manfred K. and Liao, Jun and Rätsch, Gunnar},
booktitle = {International Conference on Machine Learning},
year = {2006},
pages = {1001-1008},
doi = {10.1145/1143844.1143970},
url = {https://mlanthology.org/icml/2006/warmuth2006icml-totally/}
}