First-Order ANIL Provably Learns Representations Despite Overparametrisation
Abstract
Meta-learning methods leverage data from previous tasks to learn a new task in a sample-efficient manner. In particular, model-agnostic methods look for initialisation points from which gradient descent quickly adapts to any new task. Although it has been empirically suggested that such methods learns shared representations during pretraining, there is limited theoretical evidence of such behavior. In this direction, this work shows, in the limit of infinite tasks, first-order ANIL with a linear two-layer network successfully learns linear shared representations. This result even holds under _overparametrisation_; having a width larger than the dimension of the shared representations results in an asymptotically low-rank solution.
Cite
Text
Yüksel et al. "First-Order ANIL Provably Learns Representations Despite Overparametrisation." NeurIPS 2023 Workshops: M3L, 2023.Markdown
[Yüksel et al. "First-Order ANIL Provably Learns Representations Despite Overparametrisation." NeurIPS 2023 Workshops: M3L, 2023.](https://mlanthology.org/neuripsw/2023/yuksel2023neuripsw-firstorder/)BibTeX
@inproceedings{yuksel2023neuripsw-firstorder,
title = {{First-Order ANIL Provably Learns Representations Despite Overparametrisation}},
author = {Yüksel, Oğuz and Boursier, Etienne and Flammarion, Nicolas},
booktitle = {NeurIPS 2023 Workshops: M3L},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/yuksel2023neuripsw-firstorder/}
}