Contrastive Learning of View-invariant Representations for Facial Expressions Recognition

被引:0
|
作者
Roy, Shuvendu [1 ,2 ]
Etemad, Ali [1 ,2 ]
机构
[1] Queens Univ, Dept ECE, Kingston, ON, Canada
[2] Queens Univ, Ingenu Labs Res Inst, Kingston, ON, Canada
关键词
Affective computing; contrastive learning; expression recognition; FIELD-BASED FACE; MULTIVIEW;
D O I
10.1145/3632960
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although there has been much progress in the area of facial expression recognition (FER), most existing methods suffer when presented with images that have been captured from viewing angles that are non-frontal and substantially different from those used in the training process. In this article, we propose ViewFX, a novel view-invariant FER framework based on contrastive learning, capable of accurately classifying facial expressions regardless of the input viewing angles during inference. ViewFX learns view-invariant features of expression using a proposed self-supervised contrastive loss, which brings together different views of the same subject with a particular expression in the embedding space. We also introduce a supervised contrastive loss to push the learned view-invariant features of each expression away from other expressions. Since facial expressions are often distinguished with very subtle differences in the learned feature space, we incorporate the Barlow twins loss to reduce the redundancy and correlations of the representations in the learned representations. The proposed method is a substantial extension of our previously proposed CL-MEx, which only had a self-supervised loss. We test the proposed framework on two public multi-view facial expression recognition datasets, KDEF and DDCF. The experiments demonstrate that our approach outperforms previous works in the area and sets a new state-of-the-art for both datasets while showing considerably less sensitivity to challenging angles and the number of output labels used for training. We also perform detailed sensitivity and ablation experiments to evaluate the impact of different components of our model as well as its sensitivity to different parameters.
引用
收藏
页数:22
相关论文
共 50 条
  • [41] Dual-attention Network for View-invariant Action Recognition
    Kumie, Gedamu Alemu
    Habtie, Maregu Assefa
    Ayall, Tewodros Alemu
    Zhou, Changjun
    Liu, Huawen
    Seid, Abegaz Mohammed
    Erbad, Aiman
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (01) : 305 - 321
  • [42] Learning of view-invariant pattern recognizer with temporal context
    Inoue, K
    Urahama, K
    PATTERN RECOGNITION, 2000, 33 (10) : 1665 - 1674
  • [43] View-invariant Gesture Recognition using Nonparametric Shape Descriptor
    Wu, Xingyu
    Mao, Xia
    Chen, Lijiang
    Xue, Yuli
    Compare, Angelo
    2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, : 544 - 549
  • [44] Deeply Learned View-Invariant Features for Cross-View Action Recognition
    Kong, Yu
    Ding, Zhengming
    Li, Jun
    Fu, Yun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (06) : 3028 - 3037
  • [45] Attention Transfer (ANT) Network for View-invariant Action Recognition
    Ji, Yanli
    Xu, Feixiang
    Yang, Yang
    Xie, Ning
    Shen, Heng Tao
    Harada, Tatsuya
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 574 - 582
  • [46] View-invariant modeling and recognition of human actions using grammars
    Ogale, Abhijit S.
    Karapurkar, Alap
    Aloimonos, Yiannis
    DYNAMICAL VISION, 2007, 4358 : 115 - +
  • [47] A developmental dissociation of view-dependent and view-invariant object recognition in adolescence
    Juettner, Martin
    Mueller, Alexander
    Rentschler, Ingo
    BEHAVIOURAL BRAIN RESEARCH, 2006, 175 (02) : 420 - 424
  • [48] Development of view-invariant object recognition capability without view-invariance learning of the objects: Neuronal substrates
    Okamura, Jun-ya
    Ikejiri, Yuta
    Yamaguchi, Reona
    Wang, Gang
    Tanaka, Keiji
    NEUROSCIENCE RESEARCH, 2011, 71 : E49 - E49
  • [49] View-Invariant Action Recognition Based on Artificial Neural Networks
    Iosifidis, Alexandros
    Tefas, Anastasios
    Pitas, Ioannis
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2012, 23 (03) : 412 - 424
  • [50] View-invariant gait recognition based on kinect skeleton feature
    Sun, Jiande
    Wang, Yufei
    Li, Jing
    Wan, Wenbo
    Cheng, De
    Zhang, Huaxiang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (19) : 24909 - 24935