MMA-Net: Multi-view mixed attention mechanism for facial action unit detection

被引:3
|
作者
Shang, Ziqiao [1 ]
Du, Congju [1 ]
Li, Bingyin [1 ]
Yan, Zengqiang [1 ]
Yu, Li [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Elect Informat & Commun, Wuhan 430074, Peoples R China
基金
中国国家自然科学基金;
关键词
AU Detection; Multi-view partitioning scheme; Mixed attention mechanism; Cross-view contrastive loss;
D O I
10.1016/j.patrec.2023.06.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial action units (AU) have strong mutual correlation. How to explore fine-grained AU regional features from different dimensions while adding inter-AU correlational information is the key to accurate AU detection. In this paper, we propose a novel AU detection framework called MMA-Net based on multi view mixed attention, combining AU-regional information, co-occurrence correlational information and spatially correlational information by adopting a new AU partitioning scheme. Specifically, the proposed multi-view AU partitioning scheme first applies in both the AU co-occurrence correlational view and the facial ROI view to define the co-occurrence and spatially correlational information of AUs. Then, mixed attention, consisting of regional, channel-wise, and spatial attention, is incorporated into the encoder of MMA-Net to extract features from different dimensions. Finally, a pixel-level cross-view contrastive loss is proposed for feature enhancement by differing cross-view features for complement. Experimental results on two widely-used benchmark datasets, namely DISFA and BP4D, demonstrate the superior performance of MMA-Net against the state-of-the-art methods for AU detection.& COPY; 2023 Published by Elsevier B.V.
引用
收藏
页码:165 / 171
页数:7
相关论文
共 50 条
  • [1] Multi-view dynamic facial action unit detection
    Romero, Andres
    Leon, Juan
    Arbelaez, Pablo
    [J]. IMAGE AND VISION COMPUTING, 2022, 122
  • [2] Multi-view facial action unit detection via DenseNets and CapsNets
    Dakai Ren
    Xiangmin Wen
    Jiazhong Chen
    Yu Han
    Shiqi Zhang
    [J]. Multimedia Tools and Applications, 2022, 81 : 19377 - 19394
  • [3] Multi-view facial action unit detection via DenseNets and CapsNets
    Ren, Dakai
    Wen, Xiangmin
    Chen, Jiazhong
    Han, Yu
    Zhang, Shiqi
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (14) : 19377 - 19394
  • [4] Multi-view Surgical Video Action Detection via Mixed Global View Attention
    Schmidt, Adam
    Sharghi, Aidean
    Haugerud, Helene
    Oh, Daniel
    Mohareri, Omid
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT IV, 2021, 12904 : 626 - 635
  • [5] Multi-view facial action unit detection via deep feature enhancement
    Tang, Chuangao
    Lu, Cheng
    Zheng, Wenming
    Zong, Yuan
    Li, Sunan
    [J]. ELECTRONICS LETTERS, 2021, 57 (25) : 970 - 972
  • [6] Graph-Based Multi-Modal Multi-View Fusion for Facial Action Unit Recognition
    Chen, Jianrong
    Dey, Sujit
    [J]. IEEE ACCESS, 2024, 12 : 69310 - 69324
  • [7] Multi-view Mixed Attention for Contrastive Learning on Hypergraphs
    Lee, Jongsoo
    Chae, Dong-Kyu
    [J]. PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2543 - 2547
  • [8] Action Recognition with a Multi-View Temporal Attention Network
    Dengdi Sun
    Zhixiang Su
    Zhuanlian Ding
    Bin Luo
    [J]. Cognitive Computation, 2022, 14 : 1082 - 1095
  • [9] Action Recognition with a Multi-View Temporal Attention Network
    Sun, Dengdi
    Su, Zhixiang
    Ding, Zhuanlian
    Luo, Bin
    [J]. COGNITIVE COMPUTATION, 2022, 14 (03) : 1082 - 1095
  • [10] MMA: a multi-view and multi-modality benchmark dataset for human action recognition
    Gao, Zan
    Han, Tao-tao
    Zhang, Hua
    Xue, Yan-bing
    Xu, Guang-ping
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (22) : 29383 - 29404