UltraMAE: Multi-Modal Masked Autoencoder for Ultrasound Pre-Training

Abstract

Pre-training on a large dataset such as ImageNet followed by supervised fine-tuning has brought success in various deep learning-based tasks. However, the modalities of natural images and ultrasound images have considerable differences, making pre-training on natural images ineffective for ultrasound-related tasks. In this paper, we introduce a unified masking-based model for both ultrasound images and videos that learns better visual representation than the network with single-modality representations. This is the first large-scale generalized ultrasound pre-training network that simultaneously utilizes 100,000+ videos and images of different parts of the human anatomy such as the liver, bones, heart, thyroids, nerves, etc, making the network an effective benchmark pretrained model for any ultrasound-specific downstream tasks. We propose a novel method for ultrasound image analysis that utilizes an ultrasound-specific confidence map to guide low-level representation learning through masked feature acquisition. Our pre-trained network has demonstrated remarkable efficacy and versatility in tackling both classification and segmentation tasks across a range of ultrasound pathologies, highlighting its potential for widespread adoption and impact in the ultrasound field. In addition, we show that our pre-training model can be leveraged to learn efficiently with a small number of labeled ultrasound images.

Cite

Text

Rahman and Patel. "UltraMAE: Multi-Modal Masked Autoencoder for Ultrasound Pre-Training." Proceedings of MIDL 2024, 2024.

Markdown

[Rahman and Patel. "UltraMAE: Multi-Modal Masked Autoencoder for Ultrasound Pre-Training." Proceedings of MIDL 2024, 2024.](https://mlanthology.org/midl/2024/rahman2024midl-ultramae/)

BibTeX

@inproceedings{rahman2024midl-ultramae,
  title     = {{UltraMAE: Multi-Modal Masked Autoencoder for Ultrasound Pre-Training}},
  author    = {Rahman, Aimon and Patel, Vishal M.},
  booktitle = {Proceedings of MIDL 2024},
  year      = {2024},
  pages     = {1196-1206},
  volume    = {250},
  url       = {https://mlanthology.org/midl/2024/rahman2024midl-ultramae/}
}