GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification
Abstract
We present GazeDiff, a novel architecture that leverages radiologists\’eye gaze patterns as controls to text-to-image diffusion models for zero-shot classification. Eye-gaze patterns provide important cues during the visual exploration process; existing diffusion-based models do not harness the valuable insights derived from these patterns during image interpretation. GazeDiff utilizes a novel expert visual attention-conditioned diffusion model to generate robust medical images. This model offers more than just image generation capabilities; the density estimates derived from the gaze-guided diffusion model can effectively improve zero-shot classification performance. We show the zero-shot classification efficacy of GazeDiff on four publicly available datasets for two common pulmonary disease types, namely pneumonia, and tuberculosis.
Cite
Text
Bhattacharya and Prasanna. "GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification." Proceedings of MIDL 2024, 2024.Markdown
[Bhattacharya and Prasanna. "GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification." Proceedings of MIDL 2024, 2024.](https://mlanthology.org/midl/2024/bhattacharya2024midl-gazediff/)BibTeX
@inproceedings{bhattacharya2024midl-gazediff,
title = {{GazeDiff: A Radiologist Visual Attention Guided Diffusion Model for Zero-Shot Disease Classification}},
author = {Bhattacharya, Moinak and Prasanna, Prateek},
booktitle = {Proceedings of MIDL 2024},
year = {2024},
pages = {103-118},
volume = {250},
url = {https://mlanthology.org/midl/2024/bhattacharya2024midl-gazediff/}
}