Speed and Sparsity of Regularized Boosting
Abstract
Boosting algorithms with $l_1$ regularization are of interest because $l_1$ regularization leads to sparser composite classifiers. Moreover, Rosset et al. have shown that for separable data, standard $l_p$ regularized loss minimization results in a margin maximizing classifier in the limit as regularization is relaxed. For the case $p=1$, we extend these results by obtaining explicit convergence bounds on the regularization required to yield a margin within prescribed accuracy of the maximum achievable margin. We derive similar rates of convergence for the $\epsilon$-AdaBoost algorithm, in the process providing a new proof that $\epsilon$-AdaBoost is margin maximizing as $\epsilon$ converges to $0$. Because both of these known algorithms are computationally expensive, we introduce a new hybrid algorithm, AdaBoost+$L_1$, that combines the virtues of AdaBoost with the sparsity of $l_1$ regularization in a computationally efficient fashion. We prove that the algorithm is margin maximizing and empirically examine its performance on five datasets.
Cite
Text
Xi et al. "Speed and Sparsity of Regularized Boosting." Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, 2009.Markdown
[Xi et al. "Speed and Sparsity of Regularized Boosting." Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, 2009.](https://mlanthology.org/aistats/2009/xi2009aistats-speed/)BibTeX
@inproceedings{xi2009aistats-speed,
title = {{Speed and Sparsity of Regularized Boosting}},
author = {Xi, Yongxin and Xiang, Zhen and Ramadge, Peter and Schapire, Robert},
booktitle = {Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics},
year = {2009},
pages = {615-622},
volume = {5},
url = {https://mlanthology.org/aistats/2009/xi2009aistats-speed/}
}