Common Sense Reasoning for Deepfake Detection

被引:2
|
作者
Zhang, Yue [1 ,2 ]
Colman, Ben [2 ]
Guo, Xiao [1 ]
Shahriyari, Ali [2 ]
Bharaj, Gaurav [2 ]
机构
[1] Michigan State Univ, E Lansing, MI USA
[2] Real Defender Inc, Las Vegas, NV 89103 USA
来源
关键词
Vision and Language Model; Deepfake Detection;
D O I
10.1007/978-3-031-73223-2_22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art deepfake detection approaches rely on image-based features extracted via neural networks. While these approaches trained in a supervised manner extract likely fake features, they may fall short in representing unnatural 'non-physical' semantic facial attributes - blurry hairlines, double eyebrows, rigid eye pupils, or unnatural skin shading. However, such facial attributes are easily perceived by humans and used to discern the authenticity of an image based on human common sense. Furthermore, image-based feature extraction methods that provide visual explanations via saliency maps can be hard to interpret for humans. To address these challenges, we frame deepfake detection as a Deepfake Detection VQA (DD-VQA) task and model human intuition by providing textual explanations that describe common sense reasons for labeling an image as real or fake. We introduce a new annotated dataset and propose a Vision and Language Transformer-based framework for the DD-VQA task. We also incorporate text and image-aware feature alignment formulation to enhance multi-modal representation learning. As a result, we improve upon existing deepfake detection models by integrating our learned vision representations, which reason over common sense knowledge from the DD-VQA task. We provide extensive empirical results demonstrating that our method enhances detection performance, generalization ability, and language-based interpretability in the deepfake detection task. Our dataset is available at https://github.com/Reality-Defender/Research- DD-VQA.
引用
收藏
页码:399 / 415
页数:17
相关论文
共 50 条
  • [21] Common sense reasoning from Cyc to intelligent assistant
    Panton, Kathy
    Matuszek, Cynthia
    Lenat, Douglas
    Schneider, Dave
    Witbrock, Michael
    Siegel, Nick
    Shepard, Blake
    AMBIENT INTELLIGENCE IN EVERDAY LIFE, 2006, 3864 : 1 - 31
  • [22] COMMON-SENSE REASONING ABOUT CONSERVATION - THE ROLE OF ACTION
    MARIANI, MC
    OGBORN, J
    INTERNATIONAL JOURNAL OF SCIENCE EDUCATION, 1990, 12 (01) : 51 - 66
  • [23] Joint Common Sense and Relation Reasoning for Dense Relational Captioning
    Cao, Shan
    Liu, Weiming
    An, Gaoyun
    Ruan, Qiuqi
    PROCEEDINGS OF 2020 IEEE 15TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP 2020), 2020, : 156 - 159
  • [24] Detecting false captioning using common-sense reasoning
    Department of Electrical Engineering and Computer Science, Northwestern University, Technological Institute L359, 2145 Sheridan Road, Evanston, IL 60208, United States
    Digit. Forensic Res. Workshop, (S65-S70):
  • [25] Reactive Common Sense Reasoning for Knowledge-based HMI
    Cebulla, Michael
    FOURTH INTERNATIONAL CONFERENCE ON AUTONOMIC AND AUTONOMOUS SYSTEMS (ICAS 2008), 2008, : 41 - 46
  • [26] Cleavage politics in ordinary reasoning: How common sense divides
    Damhuis, Koen
    Westheuser, Linus
    EUROPEAN SOCIETIES, 2024, 26 (04) : 1195 - 1231
  • [27] EmoSenticSpace: A novel framework for affective common-sense reasoning
    Poria, Soujanya
    Gelbukh, Alexander
    Cambria, Erik
    Hussain, Amir
    Huang, Guang-Bin
    KNOWLEDGE-BASED SYSTEMS, 2014, 69 : 108 - 123
  • [28] Detecting false captioning using common-sense reasoning
    Lee, Sangwon
    Shamma, David A.
    Gooch, Bruce
    DIGITAL INVESTIGATION, 2006, 3 (SUPPL.) : S65 - S70
  • [29] A CONNECTION BASED APPROACH TO COMMON-SENSE TOPOLOGICAL DESCRIPTION AND REASONING
    GOTTS, NM
    GOODAY, JM
    COHN, AG
    MONIST, 1996, 79 (01): : 51 - 75
  • [30] ProtoQA: A Question Answering Dataset for Prototypical Common-Sense Reasoning
    Boratko, Michael
    Li, Xiang Lorraine
    O'Gorman, Tim
    Das, Rajarshi
    Le, Dan
    McCallum, Andrew
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1122 - 1136