Deep Multimodal Learning with Missing Modality: A Survey
Abstract
During multimodal model training and testing, certain data modalities may be absent due to sensor limitations, cost constraints, privacy concerns, or data loss, which can degrade performance. Multimodal learning techniques that explicitly account for missing modalities aim to improve robustness by enabling models to perform reliably even when certain inputs are unavailable. This survey presents the first comprehensive review of Multimodal Learning with Missing Modality (MLMM), with a focus on deep learning approaches. We outline the motivations and key distinctions between MLMM and conventional multimodal learning, provide a detailed analysis of existing methods, applications, and datasets, and conclude by highlighting open challenges and future research directions.
Cite
Text
Wu et al. "Deep Multimodal Learning with Missing Modality: A Survey." Transactions on Machine Learning Research, 2026.Markdown
[Wu et al. "Deep Multimodal Learning with Missing Modality: A Survey." Transactions on Machine Learning Research, 2026.](https://mlanthology.org/tmlr/2026/wu2026tmlr-deep/)BibTeX
@article{wu2026tmlr-deep,
title = {{Deep Multimodal Learning with Missing Modality: A Survey}},
author = {Wu, Renjie and Wang, Hu and Chen, Hsiang-Ting and Carneiro, Gustavo},
journal = {Transactions on Machine Learning Research},
year = {2026},
url = {https://mlanthology.org/tmlr/2026/wu2026tmlr-deep/}
}