Dimension-Free Exponentiated Gradient
Abstract
We present a new online learning algorithm that extends the exponentiated gradient to infinite dimensional spaces. Our analysis shows that the algorithm is implicitly able to estimate the $L_2$ norm of the unknown competitor, $U$, achieving a regret bound of the order of $O(U \log (U T+1))\sqrt{T})$, instead of the standard $O((U^2 +1) \sqrt{T})$, achievable without knowing $U$. For this analysis, we introduce novel tools for algorithms with time-varying regularizers, through the use of local smoothness. Through a lower bound, we also show that the algorithm is optimal up to $\sqrt{\log T}$ term for linear and Lipschitz losses.
Cite
Text
Orabona. "Dimension-Free Exponentiated Gradient." Neural Information Processing Systems, 2013.Markdown
[Orabona. "Dimension-Free Exponentiated Gradient." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/orabona2013neurips-dimensionfree/)BibTeX
@inproceedings{orabona2013neurips-dimensionfree,
title = {{Dimension-Free Exponentiated Gradient}},
author = {Orabona, Francesco},
booktitle = {Neural Information Processing Systems},
year = {2013},
pages = {1806-1814},
url = {https://mlanthology.org/neurips/2013/orabona2013neurips-dimensionfree/}
}