Multi-Task Network Guided Multimodal Fusion for Fake News Detection

Abstract

Fake news detection has become a hot research topic in the multimodal domain. Existing multimodal fake news detection research utilizes a series of feature fusion networks to gather useful information from different modalities of news posts. However, how to form effective cross-modal features? And how cross-modal correlations impact decision-making? These remain open questions. This paper introduces MMFND, a multi-task guided multimodal fusion framework for fake news detection , which introduces multi-task modules for feature refinement and fusion. Pairwise CLIP encoders are used to extract modality-aligned deep representations, enabling accurate measurement of cross-modal correlations. Enhancing feature fusion by weighting multimodal features with normalised cross-modal correlations. Extensive experiments on typical fake news datasets demonstrate that MMFND outperforms state-of-the-art approaches.

Cite

Text

Ma et al. "Multi-Task Network Guided Multimodal Fusion for Fake News Detection." Proceedings of the 16th Asian Conference on Machine Learning, 2024.

Markdown

[Ma et al. "Multi-Task Network Guided Multimodal Fusion for Fake News Detection." Proceedings of the 16th Asian Conference on Machine Learning, 2024.](https://mlanthology.org/acml/2024/ma2024acml-multitask/)

BibTeX

@inproceedings{ma2024acml-multitask,
  title     = {{Multi-Task Network Guided Multimodal Fusion for Fake News Detection}},
  author    = {Ma, Jinke and Zhang, Liyuan and Liu, Yong and Zhang, Wei},
  booktitle = {Proceedings of the 16th Asian Conference on Machine Learning},
  year      = {2024},
  pages     = {813-828},
  volume    = {260},
  url       = {https://mlanthology.org/acml/2024/ma2024acml-multitask/}
}