Exploring the Limits of Feature Learning in Continual Learning
Abstract
Despite the recent breakthroughs in deep learning, neural networks still struggle to learn continually in non-stationary environments, and the reasons are poorly understood. In this work, we perform an empirical study on the role of feature learning and scale on catastrophic forgetting by applying the precepts of the theory on neural networks scaling limits. We interpolate between lazy and rich training regimes, finding that the optimal amount of feature learning is modulated by task similarity. Surprisingly, our results consistently show that more feature learning increases catastrophic forgetting and that scale only helps when yielding more laziness. Supported by empirical evidence on a variety of benchmarks, our work provides the first unified understanding of the role of scale in the different training regimes and parameterizations for continual learning.
Cite
Text
Graldi et al. "Exploring the Limits of Feature Learning in Continual Learning." NeurIPS 2024 Workshops: Continual_FoMo, 2024.Markdown
[Graldi et al. "Exploring the Limits of Feature Learning in Continual Learning." NeurIPS 2024 Workshops: Continual_FoMo, 2024.](https://mlanthology.org/neuripsw/2024/graldi2024neuripsw-exploring/)BibTeX
@inproceedings{graldi2024neuripsw-exploring,
title = {{Exploring the Limits of Feature Learning in Continual Learning}},
author = {Graldi, Jacopo and Lanzillotta, Giulia and Noci, Lorenzo and Grewe, Benjamin F and Hofmann, Thomas},
booktitle = {NeurIPS 2024 Workshops: Continual_FoMo},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/graldi2024neuripsw-exploring/}
}