Attention Guided Spatio-Temporal Artifacts Extraction for Deepfake Detection

被引:1
|
作者
Wang, Zhibing [1 ,2 ]
Li, Xin [1 ,2 ]
Ni, Rongrong [1 ,2 ]
Zhao, Yao [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[2] Beijing Key Lab Adv Informat Sci & Network Techno, Beijing 100044, Peoples R China
关键词
Spatio-temporal artifacts; Attention; Deepfake detection;
D O I
10.1007/978-3-030-88013-2_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, deep-learning based model has been widely used for deepfake video detection due to its effectiveness in artifacts extraction. Most of the existing deep-learning detection methods with the attention mechanism attach more importance to the information in the spatial domain. However, the discrepancy of different frames is also important and should pay different levels of attention to temporal regions. To address this problem, this paper proposes an Attention Guided LSTM Network (AGLNet), which takes into consideration the mutual correlations in both temporal and spatial domains to effectively capture the artifacts in deepfake videos. In particular, sequential feature maps extracted from convolution and fully-connected layers of the convolutional neural network are receptively fed into the attention guided LSTM module to learn soft spatio-temporal assignment weights, which help aggregate not only detailed spatial information but also temporal information from consecutive video frames. Experiments on FaceForensics++ and Celeb-DF datasets demonstrate the superiority of the proposed AGLNet model in exploring the spatio-temporal artifacts extraction.
引用
收藏
页码:374 / 386
页数:13
相关论文
共 50 条
  • [41] Spatio-temporal relationships and video object extraction
    Deng, YN
    Manjunath, BS
    CONFERENCE RECORD OF THE THIRTY-SECOND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, VOLS 1 AND 2, 1998, : 895 - 899
  • [42] Spatio-temporal conflict detection and resolution
    Howarth R.J.
    Tsang E.P.K.
    Constraints, 1998, 3 (4) : 343 - 361
  • [43] Divide and Conquer: Question-Guided Spatio-Temporal Contextual Attention for Video Question Answering
    Jiang, Jianwen
    Chen, Ziqiang
    Lin, Haojie
    Zhao, Xibin
    Gao, Yue
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11101 - 11108
  • [44] Spatio-temporal feature points detection and extraction based on convolutional neural network
    Yang, Chaoyu
    Liu, Qian
    Liang, Yincheng
    PROCEEDINGS OF THE 2015 2ND INTERNATIONAL CONFERENCE ON ELECTRICAL, COMPUTER ENGINEERING AND ELECTRONICS (ICECEE 2015), 2015, 24 : 400 - 403
  • [45] Attention-Guided Supervised Contrastive Learning for Deepfake Detection
    Waseem, Saima
    Abu Bakar, Syed Abdul Rahman Bin Syed
    Ahmed, Bilal Ashfaq
    2024 IEEE 8TH INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING APPLICATIONS, ICSIPA, 2024,
  • [46] Spatio-temporal attention model for video content analysis
    Guironnet, M
    Guyader, N
    Pellerin, D
    Ladret, P
    2005 International Conference on Image Processing (ICIP), Vols 1-5, 2005, : 2989 - 2992
  • [47] The mechanism of visual attention is the spatio-temporal salience map
    Sperling, G
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 1996, 31 (3-4) : 3366 - 3366
  • [48] Spatio-Temporal Deformable Attention Network for Video Deblurring
    Zhang, Huicong
    Xie, Haozhe
    Yao, Hongxun
    COMPUTER VISION - ECCV 2022, PT XVI, 2022, 13676 : 581 - 596
  • [49] TRAT: Tracking by attention using spatio-temporal features
    Saribas, Hasan
    Cevikalp, Hakan
    Kopuklu, Okan
    Uzun, Bedirhan
    NEUROCOMPUTING, 2022, 492 : 150 - 161
  • [50] Spatio-Temporal Convolution-Attention Video Network
    Diba, Ali
    Sharma, Vivek
    Arzani, Mohammad. M.
    Van Gool, Luc
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 859 - 869