A Primal-Dual Convergence Analysis of Boosting

Abstract

Boosting combines weak learners into a predictor with low empirical risk. Its dual constructs a high entropy distribution upon which weak learners and training labels are uncorrelated. This manuscript studies this primal-dual relationship under a broad family of losses, including the exponential loss of AdaBoost and the logistic loss, revealing: • Weak learnability aids the whole loss family: for any ε > 0, O(ln(1/ε)) iterations suffice to produce a predictor with empirical risk ε-close to the infimum; • The circumstances granting the existence of an empirical risk minimizer may be characterized in terms of the primal and dual problems, yielding a new proof of the known rate O(ln(1/ε)); • Arbitrary instances may be decomposed into the above two, granting rate O(1/ε), with a matching lower bound provided for the logistic loss.

Cite

Text

Telgarsky. "A Primal-Dual Convergence Analysis of Boosting." Journal of Machine Learning Research, 2012.

Markdown

[Telgarsky. "A Primal-Dual Convergence Analysis of Boosting." Journal of Machine Learning Research, 2012.](https://mlanthology.org/jmlr/2012/telgarsky2012jmlr-primaldual/)

BibTeX

@article{telgarsky2012jmlr-primaldual,
  title     = {{A Primal-Dual Convergence Analysis of Boosting}},
  author    = {Telgarsky, Matus},
  journal   = {Journal of Machine Learning Research},
  year      = {2012},
  pages     = {561-606},
  volume    = {13},
  url       = {https://mlanthology.org/jmlr/2012/telgarsky2012jmlr-primaldual/}
}