Online Strongly Convex Optimization with Unknown Delays
Abstract
We investigate the problem of online convex optimization with unknown delays, in which the feedback of a decision arrives with an arbitrary delay. Previous studies have presented delayed online gradient descent (DOGD), and achieved the regret bound of O(D)\documentclass[12pt]minimal \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}-69pt \begin{document}$O(\sqrt{D})$\end{document} by only utilizing the convexity condition, where D≥T\documentclass[12pt]minimal \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}-69pt \begin{document}$D\ge T$\end{document} is the sum of delays over T rounds. In this paper, we further exploit the strong convexity to improve the regret bound. Specifically, we first propose a variant of DOGD for strongly convex functions, and establish a better regret bound of O(dlogT)\documentclass[12pt]minimal \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}-69pt \begin{document}$O(d\log T)$\end{document}, where d is the maximum delay. The essential idea is to let the learning rate decay with the total number of received feedback linearly. Furthermore, we extend the strongly convex variant of DOGD and its theoretical guarantee to the more challenging bandit setting by combining with the classical (n+1)\documentclass[12pt]minimal \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}-69pt \begin{document}$(n+1)$\end{document}-point and two-point gradient estimators, where n is the dimensionality. To the best of our knowledge, this is the first work that solves online strongly convex optimization under the general delayed setting.
Cite
Text
Wan et al. "Online Strongly Convex Optimization with Unknown Delays." Machine Learning, 2022. doi:10.1007/S10994-021-06072-WMarkdown
[Wan et al. "Online Strongly Convex Optimization with Unknown Delays." Machine Learning, 2022.](https://mlanthology.org/mlj/2022/wan2022mlj-online/) doi:10.1007/S10994-021-06072-WBibTeX
@article{wan2022mlj-online,
title = {{Online Strongly Convex Optimization with Unknown Delays}},
author = {Wan, Yuanyu and Tu, Wei-Wei and Zhang, Lijun},
journal = {Machine Learning},
year = {2022},
pages = {871-893},
doi = {10.1007/S10994-021-06072-W},
volume = {111},
url = {https://mlanthology.org/mlj/2022/wan2022mlj-online/}
}