CA-FER: Mitigating Spurious Correlation With Counterfactual Attention in Facial Expression Recognition

被引:0
|
作者
Huang, Pin-Jui [1 ]
Xie, Hongxia [2 ]
Huang, Hung-Cheng [3 ]
Shuai, Hong-Han [4 ]
Cheng, Wen-Huang [5 ]
机构
[1] Natl Yang Ming Chiao Tung Univ, Artificial Intelligence Grad Program, Hsinchu 30010, Taiwan
[2] Natl Yang Ming Chiao Tung Univ, Inst Elect, Hsinchu 30010, Taiwan
[3] Univ Calif San Diego, Comp Sci & Engn, La Jolla, CA 92093 USA
[4] Natl Yang Ming Chiao Tung Univ, Dept Elect & Comp Engn, Hsinchu 30010, Taiwan
[5] Natl Taiwan Univ NTU, Dept Comp Sci & Informat Engn, Taipei 10617, Taiwan
关键词
Feature extraction; Correlation; Face recognition; Training; Computational modeling; Image color analysis; Deep learning; Causal learning; facial expression recognition; spurious correlation;
D O I
10.1109/TAFFC.2023.3312768
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although facial expression recognition based on deep learning has become a major trend, existing methods have been found to prefer learning spurious statistical correlations and non-robust features during training. This degenerates the model's generalizability in practical situations. One of the research fields mitigating such misperception of correlations as causality is causal reasoning. In this article, we propose a learnable counterfactual attention mechanism, CA-FER, that uses causal reasoning to simultaneously optimize feature discrimination and diversity to mitigate spurious correlations in expression datasets. To the best of our knowledge, this is the first work to study the spurious correlations in facial expression recognition from a counterfactual attention perspective. Extensive experiments on a synthetic dataset and four public datasets demonstrate that our method outperforms previous methods, which shows the effectiveness and generalizability of our learnable counterfactual attention mechanism for the expression recognition task.
引用
收藏
页码:977 / 989
页数:13
相关论文
共 50 条
  • [31] Facial Expression Recognition Based on Attention Mechanism and Involution
    Guo, Jingyuan
    Dong, Yishan
    Liu, Xiaowen
    Lu, Shuhua
    Computer Engineering and Applications, 2023, 59 (23) : 95 - 103
  • [32] FRAME ATTENTION NETWORKS FOR FACIAL EXPRESSION RECOGNITION IN VIDEOS
    Meng, Debin
    Peng, Xiaojiang
    Wang, Kai
    Qiao, Yu
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3866 - 3870
  • [33] Hybrid Attention Cascade Network for Facial Expression Recognition
    Zhu, Xiaoliang
    Ye, Shihao
    Zhao, Liang
    Dai, Zhicheng
    SENSORS, 2021, 21 (06) : 1 - 16
  • [34] Histogram-based local descriptors for facial expression recognition (FER): A comprehensive study
    Turan, Cigdem
    Lam, Kin-Man
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2018, 55 : 331 - 341
  • [35] MSL-FER: MIRRORED SELF-SUPERVISED LEARNING FOR FACIAL EXPRESSION RECOGNITION
    Pan, Xiangshuai
    Liu, Weifeng
    Wang, Yanjiang
    Lu, Xiaoping
    Liu, Baodi
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1601 - 1605
  • [36] Spiking-Fer: Spiking Neural Network for Facial Expression Recognition With Event Cameras
    Barchid, Sami
    Allaert, Benjamin
    Aissaoui, Amel
    Mennesson, Jose
    Djeraba, Chaabane
    20TH INTERNATIONAL CONFERENCE ON CONTENT-BASED MULTIMEDIA INDEXING, CBMI 2023, 2023, : 1 - 7
  • [37] UA-FER: Uncertainty-aware representation learning for facial expression recognition
    Zhou, Haoliang
    Huang, Shucheng
    Xu, Yuqiao
    NEUROCOMPUTING, 2025, 621
  • [38] FER-Net: facial expression recognition using densely connected convolutional network
    Ma, Hui
    Celik, Turgay
    ELECTRONICS LETTERS, 2019, 55 (04) : 184 - +
  • [39] Facial expression recognition based on regional adaptive correlation
    Hao, Meng
    Yuan, Fei
    Li, Jing
    Sun, Yuting
    IET COMPUTER VISION, 2023, 17 (04) : 445 - 460
  • [40] Facial Expression Recognition Based on Spatial and Channel Attention Mechanisms
    Lisha Yao
    Shixiong He
    Kang Su
    Qingtong Shao
    Wireless Personal Communications, 2022, 125 : 1483 - 1500