On the Importance of Embedding Norms in Self-Supervised Learning

Abstract

Self-supervised learning (SSL) allows training data representations without a supervised signal and has become an important paradigm in machine learning. Most SSL methods employ the cosine similarity between embedding vectors and hence effectively embed data on a hypersphere. While this seemingly implies that embedding norms cannot play any role in SSL, a few recent works have suggested that embedding norms have properties related to network convergence and confidence. In this paper, we resolve this apparent contradiction and systematically establish the embedding norm’s role in SSL training. Using theoretical analysis, simulations, and experiments, we show that embedding norms (i) govern SSL convergence rates and (ii) encode network confidence, with smaller norms corresponding to unexpected samples. Additionally, we show that manipulating embedding norms can have large effects on convergence speed. Our findings demonstrate that SSL embedding norms are integral to understanding and optimizing network behavior.

Cite

Text

Draganov et al. "On the Importance of Embedding Norms in Self-Supervised Learning." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Draganov et al. "On the Importance of Embedding Norms in Self-Supervised Learning." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/draganov2025icml-importance/)

BibTeX

@inproceedings{draganov2025icml-importance,
  title     = {{On the Importance of Embedding Norms in Self-Supervised Learning}},
  author    = {Draganov, Andrew and Vadgama, Sharvaree and Damrich, Sebastian and Böhm, Jan Niklas and Maes, Lucas and Kobak, Dmitry and Bekkers, Erik J},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {14417-14438},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/draganov2025icml-importance/}
}