Artificial Constraints and Hints for Unbounded Online Learning
Abstract
We provide algorithms that guarantees regret $R_T(u)\le \tilde O(G\|u\|^3 + G(\|u\|+1)\sqrt{T})$ or $R_T(u)\le \tilde O(G\|u\|^3T^{1/3} + GT^{1/3}+ G\|u\|\sqrt{T})$ for online convex optimization with $G$-Lipschitz losses for any comparison point $u$ without prior knowledge of either $G$ or $\|u\|$. Previous algorithms dispense with the $O(\|u\|^3)$ term at the expense of knowledge of one or both of these parameters, while a lower bound shows that some additional penalty term over $G\|u\|\sqrt{T}$ is necessary. Previous penalties were \emph{exponential} while our bounds are polynomial in all quantities. Further, given a known bound $\|u\|\le D$, our same techniques allow us to design algorithms that adapt optimally to the unknown value of $\|u\|$ without requiring knowledge of $G$.
Cite
Text
Cutkosky. "Artificial Constraints and Hints for Unbounded Online Learning." Conference on Learning Theory, 2019.Markdown
[Cutkosky. "Artificial Constraints and Hints for Unbounded Online Learning." Conference on Learning Theory, 2019.](https://mlanthology.org/colt/2019/cutkosky2019colt-artificial/)BibTeX
@inproceedings{cutkosky2019colt-artificial,
title = {{Artificial Constraints and Hints for Unbounded Online Learning}},
author = {Cutkosky, Ashok},
booktitle = {Conference on Learning Theory},
year = {2019},
pages = {874-894},
volume = {99},
url = {https://mlanthology.org/colt/2019/cutkosky2019colt-artificial/}
}