Semi-IIN: Semi-Supervised Intra-Inter Modal Interaction Learning Network for Multimodal Sentiment Analysis

Abstract

Despite multimodal sentiment analysis being a fertile research ground that merits further investigation, current approaches take up high annotation cost and suffer from label ambiguity, non-amicable to high-quality labeled data acquisition. Furthermore, choosing the right interactions is essential because the significance of intra- or inter-modal interactions can differ among various samples. To this end, we propose Semi-IIN, a Semi-supervised Intra-inter modal Interaction learning Network for multimodal sentiment analysis. Semi-IIN integrates masked attention and gating mechanisms, enabling effective dynamic selection after independently capturing intra- and inter-modal interactive information. Combined with the self-training approach, Semi-IIN fully utilizes the knowledge learned from unlabeled data. Experimental results on two public datasets, MOSI and MOSEI, demonstrate the effectiveness of Semi-IIN, establishing a new state-of-the-art on several metrics.

Cite

Text

Lin et al. "Semi-IIN: Semi-Supervised Intra-Inter Modal Interaction Learning Network for Multimodal Sentiment Analysis." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I2.32131

Markdown

[Lin et al. "Semi-IIN: Semi-Supervised Intra-Inter Modal Interaction Learning Network for Multimodal Sentiment Analysis." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/lin2025aaai-semi/) doi:10.1609/AAAI.V39I2.32131

BibTeX

@inproceedings{lin2025aaai-semi,
  title     = {{Semi-IIN: Semi-Supervised Intra-Inter Modal Interaction Learning Network for Multimodal Sentiment Analysis}},
  author    = {Lin, Jinhao and Wang, Yifei and Xu, Yanwu and Liu, Qi},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {1411-1419},
  doi       = {10.1609/AAAI.V39I2.32131},
  url       = {https://mlanthology.org/aaai/2025/lin2025aaai-semi/}
}