A Simpler Unified Analysis of Budget Perceptrons

Abstract

The kernel Perceptron is an appealing online learning algorithm that has a drawback: whenever it makes an error it must increase its support set, which slows training and testing if the number of errors is large. The Forgetron and the Randomized Budget Perceptron algorithms overcome this problem by restricting the number of support vectors the Perceptron is allowed to have. These algorithms have regret bounds whose proofs are dissimilar. In this paper we propose a unified analysis of both of these algorithms by observing that the way in which they remove support vectors can be seen as types of $L_2$-regularization. By casting these algorithms as instances of online convex optimization problems and applying a variant of Zinkevich's theorem for noisy and incorrect gradient, we can bound the regret of these algorithms more easily than before. Our bounds are similar to the existing ones, but the proofs are less technical.

Cite

Text

Sutskever. "A Simpler Unified Analysis of Budget Perceptrons." International Conference on Machine Learning, 2009. doi:10.1145/1553374.1553500

Markdown

[Sutskever. "A Simpler Unified Analysis of Budget Perceptrons." International Conference on Machine Learning, 2009.](https://mlanthology.org/icml/2009/sutskever2009icml-simpler/) doi:10.1145/1553374.1553500

BibTeX

@inproceedings{sutskever2009icml-simpler,
  title     = {{A Simpler Unified Analysis of Budget Perceptrons}},
  author    = {Sutskever, Ilya},
  booktitle = {International Conference on Machine Learning},
  year      = {2009},
  pages     = {985-992},
  doi       = {10.1145/1553374.1553500},
  url       = {https://mlanthology.org/icml/2009/sutskever2009icml-simpler/}
}