Self-Supervised Learning of Phenotypic Representations from Cell Images with Weak Labels

Abstract

We propose WS-DINO as a novel framework to use weak label information in learning phenotypic representations from high-content fluorescent images of cells. Our model is based on a knowledge distillation approach with a vision transformer backbone (DINO), and we use this as a benchmark model for our study. Using WS-DINO, we fine-tuned with weak label information available in high-content microscopy screens (treatment and compound) and achieve state-of-the-art performance in not-same-compound mechanism of action prediction on the BBBC021 dataset (98%), and not-same-compound-and-batch performance (96%) using the compound as the weak label. Our method bypasses single cell cropping as a pre-processing step, and using self-attention maps we show that the model learns structurally meaningful phenotypic profiles.

Cite

Text

Cross-Zamirski et al. "Self-Supervised Learning of Phenotypic Representations from Cell Images with Weak Labels." NeurIPS 2022 Workshops: LMRL, 2022.

Markdown

[Cross-Zamirski et al. "Self-Supervised Learning of Phenotypic Representations from Cell Images with Weak Labels." NeurIPS 2022 Workshops: LMRL, 2022.](https://mlanthology.org/neuripsw/2022/crosszamirski2022neuripsw-selfsupervised/)

BibTeX

@inproceedings{crosszamirski2022neuripsw-selfsupervised,
  title     = {{Self-Supervised Learning of Phenotypic Representations from Cell Images with Weak Labels}},
  author    = {Cross-Zamirski, Jan Oscar and Williams, Guy and Mouchet, Elizabeth and Schönlieb, Carola-Bibiane and Turkki, Riku and Wang, Yinhai},
  booktitle = {NeurIPS 2022 Workshops: LMRL},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/crosszamirski2022neuripsw-selfsupervised/}
}