Multi-modal Robustness Fake News Detection with Cross-Modal and Propagation Network Contrastive Learning

被引:0
|
作者
Chen, Han [1 ,2 ]
Wang, Hairong [1 ]
Liu, Zhipeng [1 ]
Li, Yuhua [1 ]
Hu, Yifan [3 ]
Zhang, Yujing [1 ]
Shu, Kai [4 ]
Li, Ruixuan [1 ]
Yu, Philip S. [5 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, Wuhan 430074, Peoples R China
[2] Huazhong Univ Sci & Technol, Inst Artificial Intelligence, Wuhan 430074, Peoples R China
[3] Univ Sydney, Sch Comp Sci, Sydney 2006, Australia
[4] Emory Univ, Dept Comp Sci, Atlanta, GA 30322 USA
[5] Univ Illinois, Dept Comp Sci, Chicago, IL 60607 USA
基金
中国国家自然科学基金;
关键词
Contrastive learning; Multi-modal; Fake news detection; Limited labeled data; Mismatched pairs scenario;
D O I
10.1016/j.knosys.2024.112800
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Social media has transformed the landscape of news dissemination, characterized by its rapid, extensive, and diverse content, coupled with the challenge of verifying authenticity. The proliferation of multimodal news on these platforms has presented novel obstacles in detecting fake news. Existing approaches typically focus on single modalities, such as text or images, or combine text and image content or with propagation network data. However, the potential for more robust fake news detection lies in considering three modalities simultaneously. In addition, the heavy reliance on labeled data in current detection methods proves time-consuming and costly. To address these challenges, we propose a novel approach, M ulti-modal Robustness F ake News Detection with Cross-Modal and Propagation Network C ontrastive L earning (MFCL). This method integrates intrinsic features from text, images, and propagation networks, capturing essential intermodal relationships for accurate fake news detection. Contrastive learning is employed to learn intrinsic features while mitigating the issue of limited labeled data. Furthermore, we introduce image-text matching (ITM) data augmentation to ensure consistent image-text representations and employ adaptive propagation (AP) network data augmentation for high-order feature learning. We utilize contextual transformers to bolster the effectiveness of fake news detection, unveiling crucial intermodal connections in the process. Experimental results on real-world datasets demonstrate that MFCL outperforms existing methods, maintaining high accuracy and robustness even with limited labeled data and mismatched pairs. Our code is available at https://github.com/HanChen-HUST/KBSMFCL.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] ConvNet frameworks for multi-modal fake news detection
    Raj, Chahat
    Meel, Priyanka
    APPLIED INTELLIGENCE, 2021, 51 (11) : 8132 - 8148
  • [22] SpotFake: A Multi-modal Framework for Fake News Detection
    Singhal, Shivangi
    Shah, Rajiv Ratn
    Chakraborty, Tanmoy
    Kumaraguru, Ponnurangam
    Satoh, Shin'ichi
    2019 IEEE FIFTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2019), 2019, : 39 - 47
  • [23] Vae-Clip: Unveiling Deception through Cross-Modal Models and Multi-Feature Integration in Multi-Modal Fake News Detection
    Zhou, Yufeng
    Pang, Aiping
    Yu, Guang
    ELECTRONICS, 2024, 13 (15)
  • [24] Multi-Modal Sarcasm Detection with Interactive In-Modal and Cross-Modal Graphs
    Liang, Bin
    Lou, Chenwei
    Li, Xiang
    Gui, Lin
    Yang, Min
    Xu, Ruifeng
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4707 - 4715
  • [25] Multi-Modal Co-Attention Capsule Network for Fake News Detection
    Optical Memory and Neural Networks, 2024, 33 : 13 - 27
  • [26] Multi-Modal Co-Attention Capsule Network for Fake News Detection
    Yin, Chunyan
    Chen, Yongheng
    OPTICAL MEMORY AND NEURAL NETWORKS, 2024, 33 (01) : 13 - 27
  • [27] Multi-Modal Co-Attention Capsule Network for Fake News Detection
    Chunyan Yin
    Yongheng Chen
    Optical Memory and Neural Networks (Information Optics), 2024, 33 (01): : 13 - 27
  • [28] Adversarial Graph Attention Network for Multi-modal Cross-modal Retrieval
    Wu, Hongchang
    Guan, Ziyu
    Zhi, Tao
    zhao, Wei
    Xu, Cai
    Han, Hong
    Yang, Yarning
    2019 10TH IEEE INTERNATIONAL CONFERENCE ON BIG KNOWLEDGE (ICBK 2019), 2019, : 265 - 272
  • [29] Cross-modal incongruity aligning and collaborating for multi-modal sarcasm detection
    Wang, Jie
    Yang, Yan
    Jiang, Yongquan
    Ma, Minbo
    Xie, Zhuyang
    Li, Tianrui
    INFORMATION FUSION, 2024, 103
  • [30] Semantics-Enhanced Multi-Modal Fake News Detection
    Qi P.
    Cao J.
    Sheng Q.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (07): : 1456 - 1465