A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning

Abstract

Misdiagnosis causes significant harm to healthcare systems worldwide, leading to increased costs and patient risks. MedRAG is a smart multimodal healthcare copilot equipped with powerful large language model (LLM) reasoning, designed to enhance medical decision-making. It supports multiple input modalities, including non-intrusive voice monitoring, general medical queries, and electronic health records. MedRAG provides recommendations on diagnosis, treatment, medication, and follow-up questioning. Leveraging retrieval-augmented generation enhanced by knowledge graph-elicited reasoning, MedRAG retrieves and integrates critical diagnostic insights, reducing the risk of misdiagnosis. It has been evaluated on both public and private datasets, outperforming existing models and offering more specific and accurate healthcare assistance. A demonstration video of MedRAG is available at: https://www.youtube.com/watch?v=PNIBDMYRfDM. The source code is available at: https://github.com/SNOWTEAM2023/MedRAG.

Cite

Text

Zhao et al. "A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/1278

Markdown

[Zhao et al. "A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/zhao2025ijcai-smart/) doi:10.24963/IJCAI.2025/1278

BibTeX

@inproceedings{zhao2025ijcai-smart,
  title     = {{A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning}},
  author    = {Zhao, Xuejiao and Liu, Siyan and Yang, Su-Yin and Miao, Chunyan},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {11132-11136},
  doi       = {10.24963/IJCAI.2025/1278},
  url       = {https://mlanthology.org/ijcai/2025/zhao2025ijcai-smart/}
}