Worst-Case Analysis of Selective Sampling for Linear-Threshold Algorithms
Abstract
We provide a worst-case analysis of selective sampling algorithms for learning linear threshold functions. The algorithms considered in this paper are Perceptron-like algorithms, i.e., algorithms which can be effi- ciently run in any reproducing kernel Hilbert space. Our algorithms ex- ploit a simple margin-based randomized rule to decide whether to query the current label. We obtain selective sampling algorithms achieving on average the same bounds as those proven for their deterministic coun- terparts, but using much fewer labels. We complement our theoretical findings with an empirical comparison on two text categorization tasks. The outcome of these experiments is largely predicted by our theoreti- cal results: Our selective sampling algorithms tend to perform as good as the algorithms receiving the true label after each classification, while observing in practice substantially fewer labels.
Cite
Text
Cesa-bianchi et al. "Worst-Case Analysis of Selective Sampling for Linear-Threshold Algorithms." Neural Information Processing Systems, 2004.Markdown
[Cesa-bianchi et al. "Worst-Case Analysis of Selective Sampling for Linear-Threshold Algorithms." Neural Information Processing Systems, 2004.](https://mlanthology.org/neurips/2004/cesabianchi2004neurips-worstcase/)BibTeX
@inproceedings{cesabianchi2004neurips-worstcase,
title = {{Worst-Case Analysis of Selective Sampling for Linear-Threshold Algorithms}},
author = {Cesa-bianchi, Nicolò and Gentile, Claudio and Zaniboni, Luca},
booktitle = {Neural Information Processing Systems},
year = {2004},
pages = {241-248},
url = {https://mlanthology.org/neurips/2004/cesabianchi2004neurips-worstcase/}
}