Sparse landmarks for facial action unit detection using vision transformer and perceiver

被引:0
|
作者
Cakir, Duygu [1 ]
Yilmaz, Gorkem [2 ]
Arica, Nafiz [3 ]
机构
[1] Bahcesehir Univ, Fac Engn & Nat Sci, Dept Software Engn, Istanbul, Turkiye
[2] Bahcesehir Univ, Fac Engn & Nat Sci, Dept Comp Engn, Istanbul, Turkiye
[3] Piri Reis Univ, Fac Engn, Dept Informat Syst Engn, Istanbul, Turkiye
关键词
action unit detection; sparse learning; vision transformer; perceiver; RECOGNITION; PATCHES;
D O I
10.1504/IJCSE.2023.10060451
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The ability to accurately detect facial expressions, represented by facial action units (AUs), holds significant implications across diverse fields such as mental health diagnosis, security, and human-computer interaction. Although earlier approaches have made progress, the burgeoning complexity of facial actions demands more nuanced, computationally efficient techniques. This study pioneers the integration of sparse learning with vision transformer (ViT) and perceiver networks, focusing on the most active and descriptive landmarks for AU detection across both controlled (DISFA, BP4D) and in-the-wild (EmotioNet) datasets. Our novel approach, employing active landmark patches instead of the whole face, not only attains state-of-the-art performance but also uncovers insights into the differing attention mechanisms of ViT and perceiver. This fusion of techniques marks a significant advancement in facial analysis, potentially reshaping strategies in noise reduction and patch optimisation, setting a robust foundation for future research in the domain.
引用
收藏
页码:607 / 620
页数:15
相关论文
共 50 条
  • [41] Size Variant Landmark Patches for Facial Action Unit Detection
    Cakir, Duygu
    Arica, Nafiz
    7TH IEEE ANNUAL INFORMATION TECHNOLOGY, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE IEEE IEMCON-2016, 2016,
  • [42] Viewpoint Robustness of Automated Facial Action Unit Detection Systems
    Namba, Shushi
    Sato, Wataru
    Yoshikawa, Sakiko
    APPLIED SCIENCES-BASEL, 2021, 11 (23):
  • [43] Region attention and label embedding for facial action unit detection
    Song, Wei
    Li, Dong
    MULTIMEDIA SYSTEMS, 2025, 31 (02)
  • [44] Integrating Semantic and Temporal Relationships in Facial Action Unit Detection
    Li, Zhihua
    Deng, Xiang
    Li, Xiaotian
    Yin, Lijun
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 5519 - 5527
  • [45] Facial action unit detection via hybrid relational reasoning
    Zhiwen Shao
    Yong Zhou
    Bing Liu
    Hancheng Zhu
    Wen-Liang Du
    Jiaqi Zhao
    The Visual Computer, 2022, 38 : 3045 - 3057
  • [46] Facial action unit detection via hybrid relational reasoning
    Shao, Zhiwen
    Zhou, Yong
    Liu, Bing
    Zhu, Hancheng
    Du, Wen-Liang
    Zhao, Jiaqi
    VISUAL COMPUTER, 2022, 38 (9-10): : 3045 - 3057
  • [47] Selective Transfer Machine for Personalized Facial Action Unit Detection
    Chu, Wen-Sheng
    De la Torre, Fernando
    Cohn, Jeffery F.
    2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 3515 - 3522
  • [48] IdenNet: Identity-Aware Facial Action Unit Detection
    Tu, Cheng-Hao
    Yang, Chih-Yuan
    Hsu, Jane Yung-jen
    2019 14TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2019), 2019, : 238 - 245
  • [49] Learning to combine local models for Facial Action Unit detection
    Jaiswal, Shashank
    Martinez, Brais
    Valstar, Michel F.
    2015 11TH IEEE INTERNATIONAL CONFERENCE AND WORKSHOPS ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG), VOL. 6, 2015,
  • [50] Uncertain Graph Neural Networks for Facial Action Unit Detection
    Song, Tengfei
    Chen, Lisha
    Zheng, Wenming
    Ji, Qiang
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 5993 - 6001