Adaptive Gradient-Based Meta-Learning Methods

Abstract

We build a theoretical framework for designing and understanding practical meta-learning methods that integrates sophisticated formalizations of task-similarity with the extensive literature on online convex optimization and sequential prediction algorithms. Our approach enables the task-similarity to be learned adaptively, provides sharper transfer-risk bounds in the setting of statistical learning-to-learn, and leads to straightforward derivations of average-case regret bounds for efficient algorithms in settings where the task-environment changes dynamically or the tasks share a certain geometric structure. We use our theory to modify several popular meta-learning algorithms and improve their training and meta-test-time performance on standard problems in few-shot and federated learning.

Cite

Text

Khodak et al. "Adaptive Gradient-Based Meta-Learning Methods." Neural Information Processing Systems, 2019.

Markdown

[Khodak et al. "Adaptive Gradient-Based Meta-Learning Methods." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/khodak2019neurips-adaptive/)

BibTeX

@inproceedings{khodak2019neurips-adaptive,
  title     = {{Adaptive Gradient-Based Meta-Learning Methods}},
  author    = {Khodak, Mikhail and Balcan, Maria-Florina F and Talwalkar, Ameet S},
  booktitle = {Neural Information Processing Systems},
  year      = {2019},
  pages     = {5917-5928},
  url       = {https://mlanthology.org/neurips/2019/khodak2019neurips-adaptive/}
}