Learning Noisy Linear Classifiers via Adaptive and Selective Sampling
Abstract
We introduce efficient margin-based algorithms for selective sampling and filtering in binary classification tasks. Experiments on real-world textual data reveal that our algorithms perform significantly better than popular and similarly efficient competitors. Using the so-called Mammen-Tsybakov low noise condition to parametrize the instance distribution, and assuming linear label noise, we show bounds on the convergence rate to the Bayes risk of a weaker adaptive variant of our selective sampler. Our analysis reveals that, excluding logarithmic factors, the average risk of this adaptive sampler converges to the Bayes risk at rate N ^−(1+ α )(2+ α )/2(3+ α ) where N denotes the number of queried labels, and α >0 is the exponent in the low noise condition. For all $\alpha>\sqrt{3}-1\approx0.73$ this convergence rate is asymptotically faster than the rate N ^−(1+ α )/(2+ α ) achieved by the fully supervised version of the base selective sampler, which queries all labels. Moreover, for α →∞ (hard margin condition) the gap between the semi- and fully-supervised rates becomes exponential.
Cite
Text
Cavallanti et al. "Learning Noisy Linear Classifiers via Adaptive and Selective Sampling." Machine Learning, 2011. doi:10.1007/S10994-010-5191-XMarkdown
[Cavallanti et al. "Learning Noisy Linear Classifiers via Adaptive and Selective Sampling." Machine Learning, 2011.](https://mlanthology.org/mlj/2011/cavallanti2011mlj-learning/) doi:10.1007/S10994-010-5191-XBibTeX
@article{cavallanti2011mlj-learning,
title = {{Learning Noisy Linear Classifiers via Adaptive and Selective Sampling}},
author = {Cavallanti, Giovanni and Cesa-Bianchi, Nicolò and Gentile, Claudio},
journal = {Machine Learning},
year = {2011},
pages = {71-102},
doi = {10.1007/S10994-010-5191-X},
volume = {83},
url = {https://mlanthology.org/mlj/2011/cavallanti2011mlj-learning/}
}