GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification

Abstract

We present GazeDiff, a novel architecture that leverages radiologists\’eye gaze patterns as controls to text-to-image diffusion models for zero-shot classification. Eye-gaze patterns provide important cues during the visual exploration process; existing diffusion-based models do not harness the valuable insights derived from these patterns during image interpretation. GazeDiff utilizes a novel expert visual attention-conditioned diffusion model to generate robust medical images. This model offers more than just image generation capabilities; the density estimates derived from the gaze-guided diffusion model can effectively improve zero-shot classification performance. We show the zero-shot classification efficacy of GazeDiff on four publicly available datasets for two common pulmonary disease types, namely pneumonia, and tuberculosis.

Cite

Text

Bhattacharya and Prasanna. "GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification." Proceedings of MIDL 2024, 2024.

Markdown

[Bhattacharya and Prasanna. "GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification." Proceedings of MIDL 2024, 2024.](https://mlanthology.org/midl/2024/bhattacharya2024midl-gazediff/)

BibTeX

@inproceedings{bhattacharya2024midl-gazediff,
  title     = {{GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification}},
  author    = {Bhattacharya, Moinak and Prasanna, Prateek},
  booktitle = {Proceedings of MIDL 2024},
  year      = {2024},
  pages     = {103-118},
  volume    = {250},
  url       = {https://mlanthology.org/midl/2024/bhattacharya2024midl-gazediff/}
}