Human-Centred Multimodal Deep Learning Models for Chest X-Ray Diagnosis

Abstract

My thesis consists of investigating how chest X-ray images, radiologists' eye movements and patients' clinical data can be used to teach a machine how radiologists read and classify images with the goal of creating human-centric AI architectures that can (1) capture radiologists' search behavioural patterns using their eye-movements in order to improve classification in DL systems, and (2) automatically detect lesions in medical images using clinical data and eye tracking data. Heterogeneous data sources such as chest X-rays, radiologists' eye movements, and patients' clinical data can contribute to novel multimodal DL architectures that, instead of learning directly from images' pixels, will learn human classification patterns encoded in both the eye movements of the images' regions and patients' medical history. In addition to a quantitative evaluation, I plan to conduct questionnaires with expert radiologists to understand the effectiveness of the proposed multimodal DL architecture.

Cite

Text

Hsieh. "Human-Centred Multimodal Deep Learning Models for Chest X-Ray Diagnosis." International Joint Conference on Artificial Intelligence, 2023. doi:10.24963/IJCAI.2023/817

Markdown

[Hsieh. "Human-Centred Multimodal Deep Learning Models for Chest X-Ray Diagnosis." International Joint Conference on Artificial Intelligence, 2023.](https://mlanthology.org/ijcai/2023/hsieh2023ijcai-human/) doi:10.24963/IJCAI.2023/817

BibTeX

@inproceedings{hsieh2023ijcai-human,
  title     = {{Human-Centred Multimodal Deep Learning Models for Chest X-Ray Diagnosis}},
  author    = {Hsieh, Chihcheng},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {7085-7086},
  doi       = {10.24963/IJCAI.2023/817},
  url       = {https://mlanthology.org/ijcai/2023/hsieh2023ijcai-human/}
}