Facial Landmark Feature Fusion in Transfer Learning of Child Facial Expressions

被引:1
|
作者
Witherow, Megan A. [1 ]
Samad, Manar D. [2 ]
Diawara, Norou [3 ]
Iftekharuddin, Khan M. [1 ]
机构
[1] Old Dominion Univ, Dept Elect & Comp Engn, Vis Lab, Norfolk, VA 23529 USA
[2] Tennessee State Univ, Dept Comp Sci, Nashville, TN 37203 USA
[3] Old Dominion Univ, Dept Math & Stat, Norfolk, VA 23529 USA
来源
基金
美国国家科学基金会;
关键词
Facial expression recognition; transfer learning; feature fusion; facial landmarks; child facial expressions;
D O I
10.1117/12.2641898
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic classification of child facial expressions is challenging due to the scarcity of image samples with annotations. Transfer learning of deep convolutional neural networks (CNNs), pretrained on adult facial expressions, can be effectively finetuned for child facial expression classification using limited facial images of children. Recent work inspired by facial age estimation and age-invariant face recognition proposes a fusion of facial landmark features with deep representation learning to augment facial expression classification performance. We hypothesize that deep transfer learning of child facial expressions may also benefit from fusing facial landmark features. Our proposed model architecture integrates two input branches: a CNN branch for image feature extraction and a fully connected branch for processing landmark-based features. The model-derived features of these two branches are concatenated into a latent feature vector for downstream expression classification. The architecture is trained on an adult facial expression classification task. Then, the trained model is finetuned to perform child facial expression classification. The combined feature fusion and transfer learning approach is compared against multiple models: training on adult expressions only (adult baseline), child expression only (child baseline), and transfer learning from adult to child data. We also evaluate the classification performance of feature fusion without transfer learning on model performance. Training on child data, we find that feature fusion improves the 10-fold cross validation mean accuracy from 80.32% to 83.72% with similar variance. Proposed fine-tuning with landmark feature fusion of child expressions yields the best mean accuracy of 85.14%, a more than 30% improvement over the adult baseline and nearly 5% improvement over the child baseline.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Assessment method of facial palsy by amount of feature point movements at facial expressions
    Tanaka, Toshiyuki
    Nemoto, Junko
    Ohta, Manami
    Kunihiro, Takanobu
    IEEJ Transactions on Electronics, Information and Systems, 2006, 126 (03) : 368 - 375
  • [22] Automatic landmark point detection and tracking for human facial expressions
    Tie, Yun
    Guan, Ling
    EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2013,
  • [23] Recognition of Facial Expressions under Varying Conditions Using Dual-Feature Fusion
    Mahmood, Awais
    Hussain, Shariq
    Iqbal, Khalid
    Elkilani, Wail S.
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2019, 2019
  • [24] Clustering Facial Expressions with Transductive Learning
    Mamularu, Andra
    Florea, Laura
    Racoviteanu, Andrei
    Badea, Mihai
    2019 INTERNATIONAL SYMPOSIUM ON SIGNALS, CIRCUITS AND SYSTEMS (ISSCS 2019), 2019,
  • [25] FACIAL REACTIONS TO FACIAL EXPRESSIONS
    DIMBERG, U
    PSYCHOPHYSIOLOGY, 1982, 19 (06) : 643 - 647
  • [26] Intensity of the facial expressions influences the aftereffect of facial expressions
    Minemoto, Kazusa
    Yoshikawa, Sakiko
    PERCEPTION, 2015, 44 : 21 - 22
  • [27] Semi-feature Level Fusion for Bimodal Affect Regression Based on Facial and Bodily Expressions
    Zhang, Yang
    Zhang, Li
    PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS (AAMAS'15), 2015, : 1557 - 1565
  • [28] Facial reactions to schematic facial expressions
    Dimberg, U
    PSYCHOPHYSIOLOGY, 2001, 38 : S37 - S37
  • [29] Facial distinctiveness is affected by facial expressions
    Nozomi, T.
    Yamada, H.
    PERCEPTION, 2012, 41 : 112 - 113
  • [30] Fusion of facial expressions and EEG for implicit affective tagging
    Koelstra, Sander
    Patras, Ioannis
    IMAGE AND VISION COMPUTING, 2013, 31 (02) : 164 - 174