4.7 Article

MFIR: Multimodal fusion and inconsistency reasoning for explainable fake news detection

期刊

INFORMATION FUSION
卷 100, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.inffus.2023.101944

关键词

Fake news detection; Multimodal content analysis; Natural language processing; Social media analysis

向作者/读者索取更多资源

Fake news has a negative impact on our lives. Multimodal fake news detection has gained attention due to the growth of multimodal content in social media. Existing approaches focus on learning semantics and fusing modalities, but they have issues in shallow fusion and capturing inconsistent information. To address this, the MFIR model is proposed, which incorporates cross-modal infiltration fusion, multimodal inconsistent learning, and explanation reasoning for more effective and interpretable detection of fake news.
Fake news possesses a destructive and negative impact on our lives. With the rapid growth of multimodal content in social media communities, multimodal fake news detection has received increasing attention. Most existing approaches focus on learning the respective deep semantics of various modalities and integrating them by traditional fusion modes (like concatenation or addition, etc.) for improving detection performance, which has achieved a certain degree of success. However, they have two crucial issues: (1) Shallow cross -modal feature fusion, and (2) Difficulty in capturing inconsistent information. To this end, we propose Multimodal Fusion and Inconsistency Reasoning (MFIR) model to discover multimodal inconsistent semantics for explainable fake news detection. Specifically, MFIR consists of three modules: (1) Different from the traditional fusion modes, cross-modal infiltration fusion is designed, which is absorbed in continuously infiltrating and correlating another modality features into its internal semantics based on the current modality, which can well ensure the retention of the contextual semantics of the original modality; (2) Multimodal inconsistent learning not only captures the local inconsistent semantics from the perspectives of text and vision, but also integrates the two types of local semantics to discover global inconsistent semantics in multimodal content; (3) To enhance the interpretability of inconsistent semantics as evidence for users, we develop explanation reasoning layer to supplement the contextual information of inconsistent semantics, resulting in more understandable evidence semantics. Extensive experiments confirm the effectiveness of our model on three datasets and improved performance by up to 2.8%.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据