Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities

Abstract

Multimodal Sentiment Analysis (MSA) aims to infer human emotions by integrating complementary signals from diverse modalities. However, in real-world scenarios, missing modalities are common due to data corruption, sensor failure, or privacy concerns, which can significantly degrade model performance. To tackle this challenge, we propose Hyper-Modality Enhancement (HME), a novel framework that avoids explicit modality reconstruction by enriching each observed modality with semantically relevant cues retrieved from other samples. This cross-sample enhancement reduces reliance on fully observed data during training, making the method better suited to scenarios with inherently incomplete inputs. In addition, we introduce an uncertainty-aware fusion mechanism that adaptively balances original and enriched representations to improve robustness. Extensive experiments on three public benchmarks show that HME consistently outperforms state-of-the-art methods under various missing modality conditions, demonstrating its practicality in real-world MSA applications.

Cite

Text

Zhuang et al. "Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities." Advances in Neural Information Processing Systems, 2025.

Markdown

[Zhuang et al. "Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/zhuang2025neurips-hypermodality/)

BibTeX

@inproceedings{zhuang2025neurips-hypermodality,
  title     = {{Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities}},
  author    = {Zhuang, Yan and Liu, Minhao and Bai, Wei and Zhang, Yanru and Li, Wei and Deng, Jiawen and Ren, Fuji},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/zhuang2025neurips-hypermodality/}
}