An Empirical Study into Clustering of Unseen Datasets with Self-Supervised Foundation Models
Abstract
Can foundation models generalize to new datasets outside their training domain, without any retraining? Our suite of benchmarking experiments use encoders pretrained solely on ImageNet-1k with either supervised or self-supervised training techniques, clustering image datasets that were not seen during training with conventional clustering algorithms. This evaluation allows us to investigate the impact of the pretraining protocol on a model's ability to generalize outside its training domain, and explore what is natively prioritized by the model in its embeddings in a real-world scenario where novel data lacks labels. We find supervised encoders typically offer more utility than SSL encoders within the training domain, and vice-versa far outside of it, however, fine-tuned SSL encoders demonstrate the opposite trend.
Cite
Text
Lowe et al. "An Empirical Study into Clustering of Unseen Datasets with Self-Supervised Foundation Models." ICML 2024 Workshops: FM-Wild, 2024.Markdown
[Lowe et al. "An Empirical Study into Clustering of Unseen Datasets with Self-Supervised Foundation Models." ICML 2024 Workshops: FM-Wild, 2024.](https://mlanthology.org/icmlw/2024/lowe2024icmlw-empirical/)BibTeX
@inproceedings{lowe2024icmlw-empirical,
title = {{An Empirical Study into Clustering of Unseen Datasets with Self-Supervised Foundation Models}},
author = {Lowe, Scott C. and Haurum, Joakim Bruslund and Oore, Sageev and Moeslund, Thomas B. and Taylor, Graham W.},
booktitle = {ICML 2024 Workshops: FM-Wild},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/lowe2024icmlw-empirical/}
}