Dimension-Free Exponentiated Gradient

Abstract

We present a new online learning algorithm that extends the exponentiated gradient to infinite dimensional spaces. Our analysis shows that the algorithm is implicitly able to estimate the $L_2$ norm of the unknown competitor, $U$, achieving a regret bound of the order of $O(U \log (U T+1))\sqrt{T})$, instead of the standard $O((U^2 +1) \sqrt{T})$, achievable without knowing $U$. For this analysis, we introduce novel tools for algorithms with time-varying regularizers, through the use of local smoothness. Through a lower bound, we also show that the algorithm is optimal up to $\sqrt{\log T}$ term for linear and Lipschitz losses.

Cite

Text

Orabona. "Dimension-Free Exponentiated Gradient." Neural Information Processing Systems, 2013.

Markdown

[Orabona. "Dimension-Free Exponentiated Gradient." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/orabona2013neurips-dimensionfree/)

BibTeX

@inproceedings{orabona2013neurips-dimensionfree,
  title     = {{Dimension-Free Exponentiated Gradient}},
  author    = {Orabona, Francesco},
  booktitle = {Neural Information Processing Systems},
  year      = {2013},
  pages     = {1806-1814},
  url       = {https://mlanthology.org/neurips/2013/orabona2013neurips-dimensionfree/}
}