Contrasting the Landscape of Contrastive and Non-Contrastive Learning
Abstract
A lot of recent advances in unsupervised feature learning are based on designing features which are invariant under semantic data augmentations. A common way to do this is contrastive learning, which uses positive and negative samples. Some recent works however have shown promising results for non-contrastive learning, which does not require negative samples. However, the non-contrastive losses have obvious “collapsed” minima, in which the encoders output a constant feature embedding, independent of the input. A folk conjecture is that so long as these collapsed solutions are avoided, the produced feature representations should be good. In our paper, we cast doubt on this story: we show through theoretical results and controlled experiments that even on simple data models, non-contrastive losses have a preponderance of non-collapsed bad minima. Moreover, we show that the training process does not avoid these minima. Code for this work can be found at https://github.com/ashwinipokle/contrastive_landscape.
Cite
Text
Pokle et al. "Contrasting the Landscape of Contrastive and Non-Contrastive Learning." Artificial Intelligence and Statistics, 2022.Markdown
[Pokle et al. "Contrasting the Landscape of Contrastive and Non-Contrastive Learning." Artificial Intelligence and Statistics, 2022.](https://mlanthology.org/aistats/2022/pokle2022aistats-contrasting/)BibTeX
@inproceedings{pokle2022aistats-contrasting,
title = {{Contrasting the Landscape of Contrastive and Non-Contrastive Learning}},
author = {Pokle, Ashwini and Tian, Jinjin and Li, Yuchen and Risteski, Andrej},
booktitle = {Artificial Intelligence and Statistics},
year = {2022},
pages = {8592-8618},
volume = {151},
url = {https://mlanthology.org/aistats/2022/pokle2022aistats-contrasting/}
}