Contrastive Learning of View-invariant Representations for Facial Expressions Recognition

被引:0
|
作者
Roy, Shuvendu [1 ,2 ]
Etemad, Ali [1 ,2 ]
机构
[1] Queens Univ, Dept ECE, Kingston, ON, Canada
[2] Queens Univ, Ingenu Labs Res Inst, Kingston, ON, Canada
关键词
Affective computing; contrastive learning; expression recognition; FIELD-BASED FACE; MULTIVIEW;
D O I
10.1145/3632960
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although there has been much progress in the area of facial expression recognition (FER), most existing methods suffer when presented with images that have been captured from viewing angles that are non-frontal and substantially different from those used in the training process. In this article, we propose ViewFX, a novel view-invariant FER framework based on contrastive learning, capable of accurately classifying facial expressions regardless of the input viewing angles during inference. ViewFX learns view-invariant features of expression using a proposed self-supervised contrastive loss, which brings together different views of the same subject with a particular expression in the embedding space. We also introduce a supervised contrastive loss to push the learned view-invariant features of each expression away from other expressions. Since facial expressions are often distinguished with very subtle differences in the learned feature space, we incorporate the Barlow twins loss to reduce the redundancy and correlations of the representations in the learned representations. The proposed method is a substantial extension of our previously proposed CL-MEx, which only had a self-supervised loss. We test the proposed framework on two public multi-view facial expression recognition datasets, KDEF and DDCF. The experiments demonstrate that our approach outperforms previous works in the area and sets a new state-of-the-art for both datasets while showing considerably less sensitivity to challenging angles and the number of output labels used for training. We also perform detailed sensitivity and ablation experiments to evaluate the impact of different components of our model as well as its sensitivity to different parameters.
引用
收藏
页数:22
相关论文
共 50 条
  • [11] View-Invariant Representation and Recognition of Actions
    Cen Rao
    Alper Yilmaz
    Mubarak Shah
    International Journal of Computer Vision, 2002, 50 : 203 - 226
  • [12] View-invariant representation and recognition of actions
    Rao, C
    Yilmaz, A
    Shah, M
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2002, 50 (02) : 203 - 226
  • [13] Discriminative Shared Gaussian Processes for Multiview and View-Invariant Facial Expression Recognition
    Eleftheriadis, Stefanos
    Rudovic, Ognjen
    Pantic, Maja
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (01) : 189 - 204
  • [14] Evidence for view-invariant face recognition units in unfamiliar face learning
    Etchells, David B.
    Brooks, Joseph L.
    Johnston, Robert A.
    QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2017, 70 (05): : 874 - 889
  • [15] One-shot learning of view-invariant object representations in newborn chicks
    Wood, Justin N.
    Wood, Samantha M. W.
    COGNITION, 2020, 199
  • [16] View-Invariant Gait Recognition Based on Deterministic Learning and Knowledge Fusion
    Deng, Muqing
    Yang, Haonan
    Cao, Jiuwen
    Feng, Xiaoreng
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [17] Quantum Morphogenetic System In View-Invariant Recognition
    Resconi, G.
    Loo, C. K.
    Tay, N. W.
    2010 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS IJCNN 2010, 2010,
  • [18] Robust view-invariant multiscale gait recognition
    Das Choudhury, Sruti
    Tjahjadi, Tardi
    PATTERN RECOGNITION, 2015, 48 (03) : 798 - 811
  • [19] View-invariant Action Recognition in Surveillance Videos
    Zhang, Fang
    Wang, Yunhong
    Zhang, Zhaoxiang
    2011 FIRST ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2011, : 580 - 583
  • [20] View-Invariant Object Category Learning, Attention, Recognition, Search, and Scene Understanding
    Grossberg, Stephen
    IJCNN: 2009 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1- 6, 2009, : 3507 - 3509