How Classifier Features Transfer to Downstream: An Asymptotic Analysis in a Two-Layer Model

Abstract

Neural networks learn effective feature representations, which can be transferred to new tasks without additional training. While larger datasets are known to improve feature transfer, the theoretical conditions for the success of such transfer remain unclear. This work investigates feature transfer in networks trained for classification to identify the conditions that enable effective clustering in unseen classes. We first reveal that higher similarity between training and unseen distributions leads to improved Cohesion and Separability. We then show that feature expressiveness is enhanced when inputs are similar to the training classes, while the features of irrelevant inputs remain indistinguishable. We validate our analysis on synthetic and benchmark datasets, including CAR, CUB, SOP, ISC, and ImageNet. Our analysis highlights the importance of the similarity between training classes and the input distribution for successful feature transfer.

Cite

Text

Bin Yoo et al. "How Classifier Features Transfer to Downstream: An Asymptotic Analysis in a Two-Layer Model." Advances in Neural Information Processing Systems, 2025.

Markdown

[Bin Yoo et al. "How Classifier Features Transfer to Downstream: An Asymptotic Analysis in a Two-Layer Model." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/yoo2025neurips-classifier/)

BibTeX

@inproceedings{yoo2025neurips-classifier,
  title     = {{How Classifier Features Transfer to Downstream: An Asymptotic Analysis in a Two-Layer Model}},
  author    = {Bin Yoo, Hee and Lee, Sungyoon and Jang, Cheongjae and Han, Dong-Sig and Kim, Jaein and Lim, Seunghyeon and Zhang, Byoung-Tak},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/yoo2025neurips-classifier/}
}