Modulate Your Spectrum in Self-Supervised Learning

Abstract

Whitening loss offers a theoretical guarantee against feature collapse in self-supervised learning (SSL) with joint embedding architectures. Typically, it involves a hard whitening approach, transforming the embedding and applying loss to the whitened output. In this work, we introduce Spectral Transformation (ST), a framework to modulate the spectrum of embedding and to seek for functions beyond whitening that can avoid dimensional collapse. We show that whitening is a special instance of ST by definition, and our empirical investigations unveil other ST instances capable of preventing collapse. Additionally, we propose a novel ST instance named IterNorm with trace loss (INTL). Theoretical analysis confirms INTL's efficacy in preventing collapse and modulating the spectrum of embedding toward equal-eigenvalues during optimization. Our experiments on ImageNet classification and COCO object detection demonstrate INTL's potential in learning superior representations. The code is available at https://github.com/winci-ai/INTL.

Cite

Text

Weng et al. "Modulate Your Spectrum in Self-Supervised Learning." International Conference on Learning Representations, 2024.

Markdown

[Weng et al. "Modulate Your Spectrum in Self-Supervised Learning." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/weng2024iclr-modulate/)

BibTeX

@inproceedings{weng2024iclr-modulate,
  title     = {{Modulate Your Spectrum in Self-Supervised Learning}},
  author    = {Weng, Xi and Ni, Yunhao and Song, Tengwei and Luo, Jie and Anwer, Rao Muhammad and Khan, Salman and Khan, Fahad and Huang, Lei},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/weng2024iclr-modulate/}
}