Contrastive Learning of View-invariant Representations for Facial Expressions Recognition

被引:0
|
作者
Roy, Shuvendu [1 ,2 ]
Etemad, Ali [1 ,2 ]
机构
[1] Queens Univ, Dept ECE, Kingston, ON, Canada
[2] Queens Univ, Ingenu Labs Res Inst, Kingston, ON, Canada
关键词
Affective computing; contrastive learning; expression recognition; FIELD-BASED FACE; MULTIVIEW;
D O I
10.1145/3632960
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although there has been much progress in the area of facial expression recognition (FER), most existing methods suffer when presented with images that have been captured from viewing angles that are non-frontal and substantially different from those used in the training process. In this article, we propose ViewFX, a novel view-invariant FER framework based on contrastive learning, capable of accurately classifying facial expressions regardless of the input viewing angles during inference. ViewFX learns view-invariant features of expression using a proposed self-supervised contrastive loss, which brings together different views of the same subject with a particular expression in the embedding space. We also introduce a supervised contrastive loss to push the learned view-invariant features of each expression away from other expressions. Since facial expressions are often distinguished with very subtle differences in the learned feature space, we incorporate the Barlow twins loss to reduce the redundancy and correlations of the representations in the learned representations. The proposed method is a substantial extension of our previously proposed CL-MEx, which only had a self-supervised loss. We test the proposed framework on two public multi-view facial expression recognition datasets, KDEF and DDCF. The experiments demonstrate that our approach outperforms previous works in the area and sets a new state-of-the-art for both datasets while showing considerably less sensitivity to challenging angles and the number of output labels used for training. We also perform detailed sensitivity and ablation experiments to evaluate the impact of different components of our model as well as its sensitivity to different parameters.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Learning Optimised Representations for View-invariant Gait Recognition
    Jia, Ning
    Sanchez, Victor
    Li, Chang-Tsun
    2017 IEEE INTERNATIONAL JOINT CONFERENCE ON BIOMETRICS (IJCB), 2017, : 774 - 780
  • [2] Learning View-invariant Sparse Representations for Cross-view Action Recognition
    Zheng, Jingjing
    Jiang, Zhuolin
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 3176 - 3183
  • [3] View-Invariant Gait Recognition with Attentive Recurrent Learning of Partial Representations
    Sepas-Moghaddam A.
    Etemad A.
    IEEE Transactions on Biometrics, Behavior, and Identity Science, 2021, 3 (01): : 124 - 137
  • [4] Unsupervised Learning of View-invariant Action Representations
    Li, Junnan
    Wong, Yongkang
    Zhao, Qi
    Kankanhalli, Mohan S.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] View-Invariant Gait Recognition via Deterministic Learning
    Zeng, Wei
    Wang, Cong
    PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 3465 - 3472
  • [6] Hierarchically Learned View-Invariant Representations for Cross-View Action Recognition
    Liu, Yang
    Lu, Zhaoyang
    Li, Jing
    Yang, Tao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (08) : 2416 - 2430
  • [7] View-invariant gait recognition via deterministic learning
    Zeng, Wei
    Wang, Cong
    NEUROCOMPUTING, 2016, 175 : 324 - 335
  • [8] Latent Multitask Learning for View-Invariant Action Recognition
    Mahasseni, Behrooz
    Todorovic, Sinisa
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 3128 - 3135
  • [9] Joint Subspace Learning for View-Invariant Gait Recognition
    Liu, Nini
    Lu, Jiwen
    Tan, Yap-Peng
    IEEE SIGNAL PROCESSING LETTERS, 2011, 18 (07) : 431 - 434
  • [10] Focalized contrastive view-invariant learning for self-supervised skeleton-based action recognition
    Men, Qianhui
    Ho, Edmond S. L.
    Shum, Hubert P. H.
    Leung, Howard
    NEUROCOMPUTING, 2023, 537 : 198 - 209