Attention Guided Spatio-Temporal Artifacts Extraction for Deepfake Detection

被引:1
|
作者
Wang, Zhibing [1 ,2 ]
Li, Xin [1 ,2 ]
Ni, Rongrong [1 ,2 ]
Zhao, Yao [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[2] Beijing Key Lab Adv Informat Sci & Network Techno, Beijing 100044, Peoples R China
关键词
Spatio-temporal artifacts; Attention; Deepfake detection;
D O I
10.1007/978-3-030-88013-2_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, deep-learning based model has been widely used for deepfake video detection due to its effectiveness in artifacts extraction. Most of the existing deep-learning detection methods with the attention mechanism attach more importance to the information in the spatial domain. However, the discrepancy of different frames is also important and should pay different levels of attention to temporal regions. To address this problem, this paper proposes an Attention Guided LSTM Network (AGLNet), which takes into consideration the mutual correlations in both temporal and spatial domains to effectively capture the artifacts in deepfake videos. In particular, sequential feature maps extracted from convolution and fully-connected layers of the convolutional neural network are receptively fed into the attention guided LSTM module to learn soft spatio-temporal assignment weights, which help aggregate not only detailed spatial information but also temporal information from consecutive video frames. Experiments on FaceForensics++ and Celeb-DF datasets demonstrate the superiority of the proposed AGLNet model in exploring the spatio-temporal artifacts extraction.
引用
收藏
页码:374 / 386
页数:13
相关论文
共 50 条
  • [1] Video anomaly detection based on attention and efficient spatio-temporal feature extraction
    Rahimpour, Seyed Mohammad
    Kazemi, Mohammad
    Moallem, Payman
    Safayani, Mehran
    VISUAL COMPUTER, 2024, 40 (10): : 6825 - 6841
  • [2] Dynamic Difference Learning With Spatio-Temporal Correlation for Deepfake Video Detection
    Yin, Qilin
    Lu, Wei
    Li, Bin
    Huang, Jiwu
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 4046 - 4058
  • [3] Spatio-temporal fall event detection in complex scenes using attention guided LSTM
    Feng, Qi
    Gao, Chenqiang
    Wang, Lan
    Zhao, Yue
    Song, Tiecheng
    Li, Qiang
    PATTERN RECOGNITION LETTERS, 2020, 130 (130) : 242 - 249
  • [4] Gated Spatio-Temporal Attention-Guided Video Deblurring
    Suin, Maitreya
    Rajagopalan, A. N.
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7798 - 7807
  • [5] SFormer: An end-to-end spatio-temporal transformer architecture for deepfake detection
    Kingra, Staffy
    Aggarwal, Naveen
    Kaur, Nirmal
    FORENSIC SCIENCE INTERNATIONAL-DIGITAL INVESTIGATION, 2024, 51
  • [6] Spatio-Temporal Catcher: a Self-Supervised Transformer for Deepfake Video Detection
    Li, Maosen
    Li, Xurong
    Yu, Kun
    Deng, Cheng
    Huang, Heng
    Mao, Feng
    Xue, Hui
    Li, Minghao
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 8707 - 8718
  • [7] Spatio-Temporal Attention Networks for Action Recognition and Detection
    Li, Jun
    Liu, Xianglong
    Zhang, Wenxuan
    Zhang, Mingyuan
    Song, Jingkuan
    Sebe, Nicu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (11) : 2990 - 3001
  • [8] A Spatio-Temporal Isotropic Operator for the Attention-Point Extraction
    Palenichka, Roman M.
    Zaremba, Marek B.
    COMPUTER ANALYSIS OF IMAGES AND PATTERNS, PROCEEDINGS, 2009, 5702 : 318 - 325
  • [9] Spatio-temporal ontologies and attention
    University of Freiburg, Freiburg, Germany
    Spat. Cogn. Comput., 2007, 1 (13-32):
  • [10] Annoyance Models for Videos with Spatio-Temporal Artifacts
    Silva, Alexandre F.
    Farias, Mylene C. Q.
    Redi, Judith A.
    2016 EIGHTH INTERNATIONAL CONFERENCE ON QUALITY OF MULTIMEDIA EXPERIENCE (QOMEX), 2016,