Feature Learning Dynamics Under Grokking in a Sparse Parity Task

Abstract

In this paper, we analyze the phenomenon of Grokking in a sparse parity task trained with Deep Neural Networks through the lens of feature learning. In particular, we analyze the evolution of the Neural Tangent Kernel (NTK) matrix. We show that during the initial overfitting phase, the NTK’s eigenfunctions are not aligned with the predictive input features. On the other hand, at a later stage the NTK’s top eigenfunctions evolve to focus on the features of interest, which corresponds to the onset of the delayed generalization typically observed in Grokking. Our experiments can be viewed as a mechanistic interpretation of feature learning during training through the NTK eigenfunctions’ evolution.

Cite

Text

Bautiste et al. "Feature Learning Dynamics Under Grokking in a Sparse Parity Task." ICML 2024 Workshops: HiLD, 2024.

Markdown

[Bautiste et al. "Feature Learning Dynamics Under Grokking in a Sparse Parity Task." ICML 2024 Workshops: HiLD, 2024.](https://mlanthology.org/icmlw/2024/bautiste2024icmlw-feature/)

BibTeX

@inproceedings{bautiste2024icmlw-feature,
  title     = {{Feature Learning Dynamics Under Grokking in a Sparse Parity Task}},
  author    = {Bautiste, Javier Sanguino and Bachmann, Gregor and He, Bobby and Noci, Lorenzo and Hofmann, Thomas},
  booktitle = {ICML 2024 Workshops: HiLD},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/bautiste2024icmlw-feature/}
}