Agnostic Boosting

Abstract

We extend the boosting paradigm to the realistic setting of agnostic learning, that is, to a setting where the training sample is generated by an arbitrary (unknown) probability distribution over examples and labels. We define a ß-weak agnostic learner with respect to a hypothesis class F as follows: given a distribution P it outputs some hypothesis h ε F whose error is at most er _p ( F ) + β, where er _p ( F ) is the minimal error of an hypothesis from F under the distribution P (note that for some distributions the bound may exceed a half). We show a boosting algorithm that using the weak agnostic learner computes a hypothesis whose error is at most max{c_1(β)er( F )^c _2(ß), ε}, in time polynomial in 1/ε. While this generalization guarantee is significantly weaker than the one resulting from the known PAC boosting algorithms, one should note that the assumption required for β-weak agnostic learner is much weaker. In fact, an important virtue of the notion of weak agnostic learning is that in many cases such learning is achieved by efficient algorithms.

Cite

Text

Ben-David et al. "Agnostic Boosting." Annual Conference on Computational Learning Theory, 2001. doi:10.1007/3-540-44581-1_33

Markdown

[Ben-David et al. "Agnostic Boosting." Annual Conference on Computational Learning Theory, 2001.](https://mlanthology.org/colt/2001/bendavid2001colt-agnostic/) doi:10.1007/3-540-44581-1_33

BibTeX

@inproceedings{bendavid2001colt-agnostic,
  title     = {{Agnostic Boosting}},
  author    = {Ben-David, Shai and Long, Philip M. and Mansour, Yishay},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2001},
  pages     = {507-516},
  doi       = {10.1007/3-540-44581-1_33},
  url       = {https://mlanthology.org/colt/2001/bendavid2001colt-agnostic/}
}