Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities
Abstract
Multimodal Sentiment Analysis (MSA) aims to infer human emotions by integrating complementary signals from diverse modalities. However, in real-world scenarios, missing modalities are common due to data corruption, sensor failure, or privacy concerns, which can significantly degrade model performance. To tackle this challenge, we propose Hyper-Modality Enhancement (HME), a novel framework that avoids explicit modality reconstruction by enriching each observed modality with semantically relevant cues retrieved from other samples. This cross-sample enhancement reduces reliance on fully observed data during training, making the method better suited to scenarios with inherently incomplete inputs. In addition, we introduce an uncertainty-aware fusion mechanism that adaptively balances original and enriched representations to improve robustness. Extensive experiments on three public benchmarks show that HME consistently outperforms state-of-the-art methods under various missing modality conditions, demonstrating its practicality in real-world MSA applications.
Cite
Text
Zhuang et al. "Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities." Advances in Neural Information Processing Systems, 2025.Markdown
[Zhuang et al. "Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/zhuang2025neurips-hypermodality/)BibTeX
@inproceedings{zhuang2025neurips-hypermodality,
title = {{Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities}},
author = {Zhuang, Yan and Liu, Minhao and Bai, Wei and Zhang, Yanru and Li, Wei and Deng, Jiawen and Ren, Fuji},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/zhuang2025neurips-hypermodality/}
}