Multi-Task Network Guided Multimodal Fusion for Fake News Detection
Abstract
Fake news detection has become a hot research topic in the multimodal domain. Existing multimodal fake news detection research utilizes a series of feature fusion networks to gather useful information from different modalities of news posts. However, how to form effective cross-modal features? And how cross-modal correlations impact decision-making? These remain open questions. This paper introduces MMFND, a multi-task guided multimodal fusion framework for fake news detection , which introduces multi-task modules for feature refinement and fusion. Pairwise CLIP encoders are used to extract modality-aligned deep representations, enabling accurate measurement of cross-modal correlations. Enhancing feature fusion by weighting multimodal features with normalised cross-modal correlations. Extensive experiments on typical fake news datasets demonstrate that MMFND outperforms state-of-the-art approaches.
Cite
Text
Ma et al. "Multi-Task Network Guided Multimodal Fusion for Fake News Detection." Proceedings of the 16th Asian Conference on Machine Learning, 2024.Markdown
[Ma et al. "Multi-Task Network Guided Multimodal Fusion for Fake News Detection." Proceedings of the 16th Asian Conference on Machine Learning, 2024.](https://mlanthology.org/acml/2024/ma2024acml-multitask/)BibTeX
@inproceedings{ma2024acml-multitask,
title = {{Multi-Task Network Guided Multimodal Fusion for Fake News Detection}},
author = {Ma, Jinke and Zhang, Liyuan and Liu, Yong and Zhang, Wei},
booktitle = {Proceedings of the 16th Asian Conference on Machine Learning},
year = {2024},
pages = {813-828},
volume = {260},
url = {https://mlanthology.org/acml/2024/ma2024acml-multitask/}
}