Multitask Online Mirror Descent

Abstract

We introduce and analyze MT-OMD, a multitask generalization of Online Mirror Descent (OMD) which operates by sharing updates between tasks. We prove that the regret of MT-OMD is of order $\sqrt{1 + \sigma^2(N-1)}\sqrt{T}$, where $\sigma^2$ is the task variance according to the geometry induced by the regularizer, $N$ is the number of tasks, and $T$ is the time horizon. Whenever tasks are similar, that is $\sigma^2 \le 1$, our method improves upon the $\sqrt{NT}$ bound obtained by running independent OMDs on each task. We further provide a matching lower bound, and show that our multitask extensions of Online Gradient Descent and Exponentiated Gradient, two major instances of OMD, enjoy closed-form updates, making them easy to use in practice. Finally, we present experiments which support our theoretical findings.

Cite

Text

Cesa-Bianchi et al. "Multitask Online Mirror Descent." Transactions on Machine Learning Research, 2022.

Markdown

[Cesa-Bianchi et al. "Multitask Online Mirror Descent." Transactions on Machine Learning Research, 2022.](https://mlanthology.org/tmlr/2022/cesabianchi2022tmlr-multitask/)

BibTeX

@article{cesabianchi2022tmlr-multitask,
  title     = {{Multitask Online Mirror Descent}},
  author    = {Cesa-Bianchi, Nicolò and Laforgue, Pierre and Paudice, Andrea and Pontil, Massimiliano},
  journal   = {Transactions on Machine Learning Research},
  year      = {2022},
  url       = {https://mlanthology.org/tmlr/2022/cesabianchi2022tmlr-multitask/}
}