Learning Noisy Linear Classifiers via Adaptive and Selective Sampling

Abstract

We introduce efficient margin-based algorithms for selective sampling and filtering in binary classification tasks. Experiments on real-world textual data reveal that our algorithms perform significantly better than popular and similarly efficient competitors. Using the so-called Mammen-Tsybakov low noise condition to parametrize the instance distribution, and assuming linear label noise, we show bounds on the convergence rate to the Bayes risk of a weaker adaptive variant of our selective sampler. Our analysis reveals that, excluding logarithmic factors, the average risk of this adaptive sampler converges to the Bayes risk at rate N ^−(1+ α )(2+ α )/2(3+ α ) where N denotes the number of queried labels, and α >0 is the exponent in the low noise condition. For all $\alpha>\sqrt{3}-1\approx0.73$ this convergence rate is asymptotically faster than the rate N ^−(1+ α )/(2+ α ) achieved by the fully supervised version of the base selective sampler, which queries all labels. Moreover, for α →∞ (hard margin condition) the gap between the semi- and fully-supervised rates becomes exponential.

Cite

Text

Cavallanti et al. "Learning Noisy Linear Classifiers via Adaptive and Selective Sampling." Machine Learning, 2011. doi:10.1007/S10994-010-5191-X

Markdown

[Cavallanti et al. "Learning Noisy Linear Classifiers via Adaptive and Selective Sampling." Machine Learning, 2011.](https://mlanthology.org/mlj/2011/cavallanti2011mlj-learning/) doi:10.1007/S10994-010-5191-X

BibTeX

@article{cavallanti2011mlj-learning,
  title     = {{Learning Noisy Linear Classifiers via Adaptive and Selective Sampling}},
  author    = {Cavallanti, Giovanni and Cesa-Bianchi, Nicolò and Gentile, Claudio},
  journal   = {Machine Learning},
  year      = {2011},
  pages     = {71-102},
  doi       = {10.1007/S10994-010-5191-X},
  volume    = {83},
  url       = {https://mlanthology.org/mlj/2011/cavallanti2011mlj-learning/}
}