A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning
Abstract
Misdiagnosis causes significant harm to healthcare systems worldwide, leading to increased costs and patient risks. MedRAG is a smart multimodal healthcare copilot equipped with powerful large language model (LLM) reasoning, designed to enhance medical decision-making. It supports multiple input modalities, including non-intrusive voice monitoring, general medical queries, and electronic health records. MedRAG provides recommendations on diagnosis, treatment, medication, and follow-up questioning. Leveraging retrieval-augmented generation enhanced by knowledge graph-elicited reasoning, MedRAG retrieves and integrates critical diagnostic insights, reducing the risk of misdiagnosis. It has been evaluated on both public and private datasets, outperforming existing models and offering more specific and accurate healthcare assistance. A demonstration video of MedRAG is available at: https://www.youtube.com/watch?v=PNIBDMYRfDM. The source code is available at: https://github.com/SNOWTEAM2023/MedRAG.
Cite
Text
Zhao et al. "A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/1278Markdown
[Zhao et al. "A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/zhao2025ijcai-smart/) doi:10.24963/IJCAI.2025/1278BibTeX
@inproceedings{zhao2025ijcai-smart,
title = {{A Smart Multimodal Healthcare Copilot with Powerful LLM Reasoning}},
author = {Zhao, Xuejiao and Liu, Siyan and Yang, Su-Yin and Miao, Chunyan},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2025},
pages = {11132-11136},
doi = {10.24963/IJCAI.2025/1278},
url = {https://mlanthology.org/ijcai/2025/zhao2025ijcai-smart/}
}