Squeezing Performance from Pathology Foundation Models with Chained Hyperparameter Searches

Abstract

Self-supervised learning (SSL) is perfectly suited for applications in digital pathology due to the scarcity of labeled data. Over the past years, many academic and industrial labs have published pathology foundation models, claiming `state-of-the-art' performance due to improvements in architecture, methodology, and/or training data. In this paper, we demonstrate that simply tuning the hyperparameters of popular SSL method DINOv2, using a relatively small dataset, leads to similar or superior performance. Specifically, we conduct three successive hyperparameter searches, iteratively increasing either dataset or model size while narrowing the hyperparameter search space and carrying over promising hyperparameters. Overall, this preliminary study demonstrates the importance of hyperparameter tuning in this domain and proposes straight-forward strategies to improve foundation models with additional compute and data.

Cite

Text

Cappadona et al. "Squeezing Performance from Pathology Foundation Models with Chained Hyperparameter Searches." NeurIPS 2024 Workshops: SSL, 2024.

Markdown

[Cappadona et al. "Squeezing Performance from Pathology Foundation Models with Chained Hyperparameter Searches." NeurIPS 2024 Workshops: SSL, 2024.](https://mlanthology.org/neuripsw/2024/cappadona2024neuripsw-squeezing/)

BibTeX

@inproceedings{cappadona2024neuripsw-squeezing,
  title     = {{Squeezing Performance from Pathology Foundation Models with Chained Hyperparameter Searches}},
  author    = {Cappadona, Joseph and Zeng, Ken Gary and Fernandez-Granda, Carlos and Witowski, Jan and LeCun, Yann and Geras, Krzysztof J.},
  booktitle = {NeurIPS 2024 Workshops: SSL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/cappadona2024neuripsw-squeezing/}
}