Online Linear Optimization via Smoothing

Abstract

We present a new optimization-theoretic approach to analyzing Follow-the-Leader style algorithms, particularly in the setting where perturbations are used as a tool for regularization. We show that adding a strongly convex penalty function to the decision rule and adding stochastic perturbations to data correspond to deterministic and stochastic smoothing operations, respectively. We establish an equivalence between “Follow the Regularized Leader” and “Follow the Perturbed Leader” up to the smoothness properties. This intuition leads to a new generic analysis framework that recovers and improves the previous known regret bounds of the class of algorithms commonly known as Follow the Perturbed Leader.

Cite

Text

Abernethy et al. "Online Linear Optimization via Smoothing." Annual Conference on Computational Learning Theory, 2014.

Markdown

[Abernethy et al. "Online Linear Optimization via Smoothing." Annual Conference on Computational Learning Theory, 2014.](https://mlanthology.org/colt/2014/abernethy2014colt-online/)

BibTeX

@inproceedings{abernethy2014colt-online,
  title     = {{Online Linear Optimization via Smoothing}},
  author    = {Abernethy, Jacob D. and Lee, Chansoo and Sinha, Abhinav and Tewari, Ambuj},
  booktitle = {Annual Conference on Computational Learning Theory},
  year      = {2014},
  pages     = {807-823},
  url       = {https://mlanthology.org/colt/2014/abernethy2014colt-online/}
}