Feature Collapse

Abstract

We formalize and study a phenomenon called *feature collapse* that makes precise the intuitive idea that entities playing a similar role in a learning task receive similar representations. As feature collapse requires a notion of task, we leverage a synthetic task in which a learner must classify `sentences' constituted of $L$ tokens. We start by showing experimentally that feature collapse goes hand in hand with generalization. We then prove that, in the large sample limit, distinct tokens that play identical roles in the task receive identical local feature representations in the first layer of the network. This analysis shows that a neural network trained on this task provably learns interpretable and meaningful representations in its first layer.

Cite

Text

Laurent et al. "Feature Collapse." International Conference on Learning Representations, 2024.

Markdown

[Laurent et al. "Feature Collapse." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/laurent2024iclr-feature/)

BibTeX

@inproceedings{laurent2024iclr-feature,
  title     = {{Feature Collapse}},
  author    = {Laurent, Thomas and von Brecht, James and Bresson, Xavier},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/laurent2024iclr-feature/}
}