Online (and Offline) on an Even Tighter Budget
Abstract
We develop a fast online kernel algorithm for classification which can be viewed as an improvement over the one suggested by (Crammer, Kandola and Singer, 2004), titled ”Online Classificaton on a Budget”. In that previous work, the authors introduced an on-the-fly compression of the number of examples used in the prediction function using the size of the margin as a quality measure. Although displaying impressive results on relatively noise-free data we show how their algorithm is susceptible in noisy problems. Utilizing a new quality measure for an included example, namely the error induced on a selected subset of the training data, we gain improved compression rates and insensitivity to noise over the existing approach. Our method is also extendable to the batch training mode case.
Cite
Text
Weston et al. "Online (and Offline) on an Even Tighter Budget." Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics, 2005.Markdown
[Weston et al. "Online (and Offline) on an Even Tighter Budget." Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics, 2005.](https://mlanthology.org/aistats/2005/weston2005aistats-online/)BibTeX
@inproceedings{weston2005aistats-online,
title = {{Online (and Offline) on an Even Tighter Budget}},
author = {Weston, Jason and Bordes, Antoine and Bottou, Leon},
booktitle = {Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics},
year = {2005},
pages = {413-420},
volume = {R5},
url = {https://mlanthology.org/aistats/2005/weston2005aistats-online/}
}