MedualTime: A Dual-Adapter Language Model for Medical Time Series-Text Multimodal Learning
Abstract
The recent rapid advancements in language models (LMs) have garnered attention in medical time series-text multimodal learning. However, existing contrastive learning-based and prompt-based LM approaches tend to be biased, often assigning a primary role to time series modality while treating text modality as secondary. We classify these approaches under a temporal-primary paradigm, which may overlook the unique and critical task-relevant information embedded in text modality like clinical reports, thus failing to fully leverage mutual benefits and complementarity of different modalities. To fill this gap, we propose a novel textual-temporal multimodal learning paradigm that enables either modality to serve as the primary while being enhanced by the other, thereby effectively capturing modality-specific information and fostering cross-modal interaction. In specific, we design MedualTime, a language model composed of dual adapters to implement temporal-primary and textual-primary modeling simultaneously. Within each adapter, lightweight adaptation tokens are injected into the top layers of LM to encourage high-level modality fusion. The shared LM pipeline by dual adapters not only achieves adapter alignment but also enables efficient fine-tuning, reducing computational resources. Empirically, MedualTime demonstrates superior performance on medical data, achieving notable improvements of 8% accuracy and 12% F1 in supervised settings. Furthermore, MedualTime's transferability is validated by few-shot transfer experiments from coarse-grained to fine-grained medical data.
Cite
Text
Ye et al. "MedualTime: A Dual-Adapter Language Model for Medical Time Series-Text Multimodal Learning." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/880Markdown
[Ye et al. "MedualTime: A Dual-Adapter Language Model for Medical Time Series-Text Multimodal Learning." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/ye2025ijcai-medualtime/) doi:10.24963/IJCAI.2025/880BibTeX
@inproceedings{ye2025ijcai-medualtime,
title = {{MedualTime: A Dual-Adapter Language Model for Medical Time Series-Text Multimodal Learning}},
author = {Ye, Jiexia and Zhang, Weiqi and Li, Ziyue and Li, Jia and Zhao, Meng and Tsung, Fugee},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2025},
pages = {7913-7921},
doi = {10.24963/IJCAI.2025/880},
url = {https://mlanthology.org/ijcai/2025/ye2025ijcai-medualtime/}
}