Benchmarking Self-Supervised Learning for Single-Cell Data
Abstract
Self-supervised learning (SSL) has emerged as a powerful approach for learning biologically meaningful representations of single-cell data. To establish best practices in this domain, we present a comprehensive benchmark evaluating eight SSL methods across three downstream tasks and eight datasets, with various data augmentation strategies. Our results demonstrate that SimCLR and VICReg consistently outperform other methods across different tasks. Furthermore, we identify random masking as the most effective augmentation technique. This benchmark provides valuable insights into the application of SSL to single-cell data analysis, bridging the gap between SSL and single-cell biology.
Cite
Text
Toma et al. "Benchmarking Self-Supervised Learning for Single-Cell Data." NeurIPS 2024 Workshops: SSL, 2024.Markdown
[Toma et al. "Benchmarking Self-Supervised Learning for Single-Cell Data." NeurIPS 2024 Workshops: SSL, 2024.](https://mlanthology.org/neuripsw/2024/toma2024neuripsw-benchmarking/)BibTeX
@inproceedings{toma2024neuripsw-benchmarking,
title = {{Benchmarking Self-Supervised Learning for Single-Cell Data}},
author = {Toma, Philip and Ovcharenko, Olga and Daunhawer, Imant and Vogt, Julia E and Barkmann, Florian and Boeva, Valentina},
booktitle = {NeurIPS 2024 Workshops: SSL},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/toma2024neuripsw-benchmarking/}
}