Adaptive Regularization of Weight Vectors
Abstract
We present AROW, a new online learning algorithm that combines several properties of successful : large margin training, confidence weighting, and the capacity to handle non-separable data. AROW performs adaptive regularization of the prediction function upon seeing each new instance, allowing it to perform especially well in the presence of label noise. We derive a mistake bound, similar in form to the second order perceptron bound, which does not assume separability. We also relate our algorithm to recent confidence-weighted online learning techniques and empirically show that AROW achieves state-of-the-art performance and notable robustness in the case of non-separable data.
Cite
Text
Crammer et al. "Adaptive Regularization of Weight Vectors." Neural Information Processing Systems, 2009.Markdown
[Crammer et al. "Adaptive Regularization of Weight Vectors." Neural Information Processing Systems, 2009.](https://mlanthology.org/neurips/2009/crammer2009neurips-adaptive/)BibTeX
@inproceedings{crammer2009neurips-adaptive,
title = {{Adaptive Regularization of Weight Vectors}},
author = {Crammer, Koby and Kulesza, Alex and Dredze, Mark},
booktitle = {Neural Information Processing Systems},
year = {2009},
pages = {414-422},
url = {https://mlanthology.org/neurips/2009/crammer2009neurips-adaptive/}
}