ECENet: Explainable and Context-Enhanced Network for Multi-modal Fact Verification

被引:1
|
作者
Zhang, Fanrui [1 ]
Liu, Jiawei [1 ]
Zhang, Qiang [1 ]
Sun, Esther [2 ]
Xie, Jingyi [1 ]
Zha, Zheng-Jun [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Anhui, Peoples R China
[2] Univ Toronto, Toronto, ON, Canada
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Muti-modal fact verification; Attention mechanism; Deep reinforcement learning; Interpretability;
D O I
10.1145/3581783.3612183
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, falsified claims incorporating both text and images have been disseminated more effectively than those containing text alone, raising significant concerns for multi-modal fact verification. Existing research makes contributions to multi-modal feature extraction and interaction, but fails to fully utilize and enhance the valuable and intricate semantic relationships between distinct features. Moreover, most detectors merely provide a single outcome judgment and lack an inference process or explanation. Taking these factors into account, we propose a novel Explainable and Context-Enhanced Network (ECENet) for multi-modal fact verification, making the first attempt to integrate multi-clue feature extraction, multi-level feature reasoning, and justification (explanation) generation within a unified framework. Specifically, we propose an Improved Coarse- and Fine-grained Attention Network, equipped with two types of level-grained attention mechanisms, to facilitate a comprehensive understanding of contextual information. Furthermore, we propose a novel justification generation module via deep reinforcement learning that does not require additional labels. In this module, a sentence extractor agent measures the importance between the query claim and all document sentences at each time step, selecting a suitable amount of high-scoring sentences to be rewritten as the explanation of the model. Extensive experiments demonstrate the effectiveness of the proposed method.
引用
收藏
页码:1231 / 1240
页数:10
相关论文
共 50 条
  • [1] Multi-modal Knowledge-aware Reinforcement Learning Network for Explainable Recommendation
    Tao, Shaohua
    Qiu, Runhe
    Ping, Yuan
    Ma, Hui
    KNOWLEDGE-BASED SYSTEMS, 2021, 227
  • [2] Context-enhanced authentication for infrastructureless network environments
    Wishart, Ryan
    Indulska, Jadwiga
    Portmann, Marius
    Sutton, Peter
    UBIQUITOUS INTELLIGENCE AND COMPUTING, PROCEEDINGS, 2006, 4159 : 924 - 935
  • [3] Multi-modal Identity Verification Based on Improved BP Neural Network
    Luan Fang-jun
    Li Kai
    Ma Si-liang
    PROCEEDINGS OF THE 2009 2ND INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, VOLS 1-9, 2009, : 2153 - 2157
  • [4] Multi-modal network Protocols
    Balan, RK
    Akella, A
    Seshan, S
    ACM SIGCOMM COMPUTER COMMUNICATION REVIEW, 2002, 32 (01) : 60 - 60
  • [5] MCDAN: A Multi-Scale Context-Enhanced Dynamic Attention Network for Diffusion Prediction
    Wang, Xiaowen
    Wang, Lanjun
    Su, Yuting
    Zhang, Yongdong
    Liu, An-An
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7850 - 7862
  • [6] Adaptive Context-Aware Multi-Modal Network for Depth Completion
    Zhao, Shanshan
    Gong, Mingming
    Fu, Huan
    Tao, Dacheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 5264 - 5276
  • [7] Noise compensation in a multi-modal verification system
    Sanderson, C
    Paliwal, KK
    2001 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-VI, PROCEEDINGS: VOL I: SPEECH PROCESSING 1; VOL II: SPEECH PROCESSING 2 IND TECHNOL TRACK DESIGN & IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS NEURALNETWORKS FOR SIGNAL PROCESSING; VOL III: IMAGE & MULTIDIMENSIONAL SIGNAL PROCESSING MULTIMEDIA SIGNAL PROCESSING, 2001, : 157 - 160
  • [8] Multi-modal Knowledge-aware Hierarchical Attention Network for Explainable Medical Question Answering
    Zhang, Yingying
    Qian, Shengsheng
    Fang, Quan
    Xu, Changsheng
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 1089 - 1097
  • [9] An enhanced multi-modal brain graph network for classifying neuropsychiatric disorders
    Liu, Liangliang
    Wang, Yu-Ping
    Wang, Yi
    Zhang, Pei
    Xiong, Shufeng
    MEDICAL IMAGE ANALYSIS, 2022, 81
  • [10] LCEMH: Label Correlation Enhanced Multi-modal Hashing for efficient multi-modal retrieval
    Zheng, Chaoqun
    Zhu, Lei
    Zhang, Zheng
    Duan, Wenjun
    Lu, Wenpeng
    INFORMATION SCIENCES, 2024, 659