Facial Action Unit Recognition Based on Self-Attention Spatiotemporal Fusion

被引:0
|
作者
Liang, Chaolei [1 ]
Zou, Wei [1 ]
Hu, Danfeng [1 ]
Wang, JiaJun [1 ]
机构
[1] Sch Elect & Informat Engn, Suzhou, Peoples R China
关键词
Facial action unit; Graph Convolutional Neural Network; Attention mechanism; Spatio-temporal relation;
D O I
10.1145/3670105.3670210
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Facial Action Units (AUs) serve as a precise descriptor of facial expressions, revealing an individual's psychological and mental state. Therefore, AU detection plays important roles in facial expression recognition. Existing methods often focus on extracting intra-frame information while pay less attention to inter-frame feature changes. To address this issue, this paper proposes a self-attention spatiotemporal fusion method (SAtt-STPN). In this method, a feature extractor (AFE) is specifically designed to extract uniform feature information from both strongly and weakly correlated regions. A spatiotemporal perception (STP) module is specifically designed to capture temporal information for each AU through mutually-driven independent branches in both spatial and temporal dimensions while a graph convolutional network is adopted to model intra-frame AU relationships (ARM). Ultimately, intra-frame and inter-frame information are weighted and fused for classification. Experimental results on two public datasets (BP4D and DISFA) show that the our proposed SAtt-STPN outperforms state-of-the-art methods in facial AU detection.
引用
收藏
页码:600 / 605
页数:6
相关论文
共 50 条
  • [1] SAT-Net: Self-Attention and Temporal Fusion for Facial Action Unit Detection
    Li, Zhihua
    Zhang, Zheng
    Yin, Lijun
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 5036 - 5043
  • [2] Spatiotemporal Self-attention Modeling with Temporal Patch Shift for Action Recognition
    Xiang, Wangmeng
    Li, Chao
    Wang, Biao
    Wei, Xihan
    Hua, Xian-Sheng
    Zhang, Lei
    COMPUTER VISION - ECCV 2022, PT III, 2022, 13663 : 627 - 644
  • [3] An Effective Video Transformer With Synchronized Spatiotemporal and Spatial Self-Attention for Action Recognition
    Alfasly, Saghir
    Chui, Charles K.
    Jiang, Qingtang
    Lu, Jian
    Xu, Chen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 2496 - 2509
  • [4] Facial Action Unit Detection by Adaptively Constraining Self-Attention and Causally Deconfounding Sample
    Shao, Zhiwen
    Zhu, Hancheng
    Zhou, Yong
    Xiang, Xiang
    Liu, Bing
    Yao, Rui
    Ma, Lizhuang
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, : 1711 - 1726
  • [5] A visual self-attention network for facial expression recognition
    Yu, Naigong
    Bai, Deguo
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [6] Self-Attention Networks For Motion Posture Recognition Based On Data Fusion
    Ji, Zhihao
    Xie, Qiang
    4TH INTERNATIONAL CONFERENCE ON INFORMATICS ENGINEERING AND INFORMATION SCIENCE (ICIEIS2021), 2022, 12161
  • [7] An efficient self-attention network for skeleton-based action recognition
    Xiaofei Qin
    Rui Cai
    Jiabin Yu
    Changxiang He
    Xuedian Zhang
    Scientific Reports, 12 (1)
  • [8] Self-Attention Network for Skeleton-based Human Action Recognition
    Cho, Sangwoo
    Maqbool, Muhammad Hasan
    Liu, Fei
    Foroosh, Hassan
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 624 - 633
  • [9] An efficient self-attention network for skeleton-based action recognition
    Qin, Xiaofei
    Cai, Rui
    Yu, Jiabin
    He, Changxiang
    Zhang, Xuedian
    SCIENTIFIC REPORTS, 2022, 12 (01):
  • [10] Global Positional Self-Attention for Skeleton-Based Action Recognition
    Kim, Jaehwan
    Lee, Junsuk
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 3355 - 3361