Adaptive Regularization of Weight Vectors

Abstract

We present AROW, a new online learning algorithm that combines several properties of successful : large margin training, confidence weighting, and the capacity to handle non-separable data. AROW performs adaptive regularization of the prediction function upon seeing each new instance, allowing it to perform especially well in the presence of label noise. We derive a mistake bound, similar in form to the second order perceptron bound, which does not assume separability. We also relate our algorithm to recent confidence-weighted online learning techniques and empirically show that AROW achieves state-of-the-art performance and notable robustness in the case of non-separable data.

Cite

Text

Crammer et al. "Adaptive Regularization of Weight Vectors." Neural Information Processing Systems, 2009.

Markdown

[Crammer et al. "Adaptive Regularization of Weight Vectors." Neural Information Processing Systems, 2009.](https://mlanthology.org/neurips/2009/crammer2009neurips-adaptive/)

BibTeX

@inproceedings{crammer2009neurips-adaptive,
  title     = {{Adaptive Regularization of Weight Vectors}},
  author    = {Crammer, Koby and Kulesza, Alex and Dredze, Mark},
  booktitle = {Neural Information Processing Systems},
  year      = {2009},
  pages     = {414-422},
  url       = {https://mlanthology.org/neurips/2009/crammer2009neurips-adaptive/}
}