From Lazy to Rich: Exact Learning Dynamics in Deep Linear Networks
Abstract
Biological and artificial neural networks create internal representations for complex tasks. In artificial networks, the ability to form task-specific representations is shaped by datasets, architectures, initialization strategies, and optimization algorithms. Previous studies show that different initializations lead to either a lazy regime, where representations stay static, or a rich regime, where they evolve dynamically. This work examines how initialization affects learning dynamics in deep linear networks, deriving exact solutions for $\lambda$-balanced initializations, which reflect the weight scaling across layers. These solutions explain how representations and the Neural Tangent Kernel evolve from rich to lazy regimes, with implications for continual, reversal, and transfer learning in neuroscience and practical applications.
Cite
Text
Dominé et al. "From Lazy to Rich: Exact Learning Dynamics in Deep Linear Networks." NeurIPS 2024 Workshops: M3L, 2024.Markdown
[Dominé et al. "From Lazy to Rich: Exact Learning Dynamics in Deep Linear Networks." NeurIPS 2024 Workshops: M3L, 2024.](https://mlanthology.org/neuripsw/2024/domine2024neuripsw-lazy/)BibTeX
@inproceedings{domine2024neuripsw-lazy,
title = {{From Lazy to Rich: Exact Learning Dynamics in Deep Linear Networks}},
author = {Dominé, Clémentine Carla Juliette and Anguita, Nicolas and Proca, Alexandra Maria and Braun, Lukas and Kunin, Daniel and Mediano, Pedro A. M. and Saxe, Andrew M},
booktitle = {NeurIPS 2024 Workshops: M3L},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/domine2024neuripsw-lazy/}
}