Regret Bounds for Lifelong Learning

Abstract

We consider the problem of transfer learning in an online setting. Different tasks are presented sequentially and processed by a within-task algorithm. We propose a lifelong learning strategy which refines the underlying data representation used by the within-task algorithm, thereby transferring information from one task to the next. We show that when the within-task algorithm comes with some regret bound, our strategy inherits this good property. Our bounds are in expectation for a general loss function, and uniform for a convex loss. We discuss applications to dictionary learning and finite set of predictors. In the latter case, we improve previous $O(1/\sqrt{m})$ bounds to $O(1/m)$ where $m$ is the per task sample size.

Cite

Text

Alquier et al. "Regret Bounds for Lifelong Learning." International Conference on Artificial Intelligence and Statistics, 2017.

Markdown

[Alquier et al. "Regret Bounds for Lifelong Learning." International Conference on Artificial Intelligence and Statistics, 2017.](https://mlanthology.org/aistats/2017/alquier2017aistats-regret/)

BibTeX

@inproceedings{alquier2017aistats-regret,
  title     = {{Regret Bounds for Lifelong Learning}},
  author    = {Alquier, Pierre and Mai, The Tien and Pontil, Massimiliano},
  booktitle = {International Conference on Artificial Intelligence and Statistics},
  year      = {2017},
  pages     = {261-269},
  url       = {https://mlanthology.org/aistats/2017/alquier2017aistats-regret/}
}