Loss Relaxation Strategy for Noisy Facial Video-based Automatic Depression Recognition

被引:0
|
作者
Song S. [1 ]
Luo Y. [2 ]
Tumer T. [3 ]
Fu C. [4 ]
Valstar M. [5 ]
Gunes H. [6 ]
机构
[1] University of Cambridge, Cambridge, Leicester
[2] Imperial College London, London
[3] Middle East Technical University, Ankara
[4] Northeastern University, Qinhuangdao
[5] University of Nottingham, Nottingham
[6] University of Cambridge, Cambridge
来源
基金
英国工程与自然科学研究理事会;
关键词
Depression classification; depression severity estimation; facial video analysis; loss relaxation strategy; noisy data and annotation;
D O I
10.1145/3648696
中图分类号
学科分类号
摘要
Automatic depression analysis has been widely investigated on face videos that have been carefully collected and annotated in lab conditions. However, videos collected under real-world conditions may suffer from various types of noise due to challenging data acquisition conditions and lack of annotators. Although deep learning (DL) models frequently show excellent depression analysis performances on datasets collected in controlled lab conditions, such noise may degrade their generalization abilities for real-world depression analysis tasks. In this article, we uncovered that noisy facial data and annotations consistently change the distribution of training losses for facial depression DL models; i.e., noisy data-label pairs cause larger loss values compared to clean data-label pairs. Since different loss functions could be applied depending on the employed model and task, we propose a generic loss function relaxation strategy that can jointly reduce the negative impact of various noisy data and annotation problems occurring in both classification and regression loss functions for face video-based depression analysis, where the parameters of the proposed strategy can be automatically adapted during depression model training. The experimental results on 25 different artificially created noisy depression conditions (i.e., five noise types with five different noise levels) show that our loss relaxation strategy can clearly enhance both classification and regression loss functions, enabling the generation of superior face video-based depression analysis models under almost all noisy conditions. Our approach is robust to its main variable settings and can adaptively and automatically obtain its parameters during training. © 2024 Copyright held by the owner/author(s).
引用
下载
收藏
相关论文
共 50 条
  • [1] An Automatic Framework for Textured 3D Video-Based Facial Expression Recognition
    Hayat, Munawar
    Bennamoun, Mohammed
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2014, 5 (03) : 301 - 313
  • [2] A Depth Video-based Facial Expression Recognition System
    Uddin, Md. Zia
    IETE TECHNICAL REVIEW, 2012, 29 (02) : 169 - 177
  • [3] An automatic system for unconstrained video-based face recognition
    Zheng J.
    Ranjan R.
    Chen C.-H.
    Chen J.-C.
    Castillo C.D.
    Chellappa R.
    IEEE Transactions on Biometrics, Behavior, and Identity Science, 2020, 2 (03): : 194 - 209
  • [4] A Video-Based Facial Motion Tracking and Expression Recognition System
    Jun Yu
    Zengfu Wang
    Multimedia Tools and Applications, 2017, 76 : 14653 - 14672
  • [5] An Ensemble of VGG Networks for Video-Based Facial Expression Recognition
    Jiao, Zirui
    Qiao, Fengchun
    Yao, Naiming
    Li, Zhihao
    Chen, Hui
    Wang, Hongan
    2018 FIRST ASIAN CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII ASIA), 2018,
  • [6] A Video-Based Facial Motion Tracking and Expression Recognition System
    Yu, Jun
    Wang, Zengfu
    MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (13) : 14653 - 14672
  • [7] Video-based facial expression recognition by removing the style variations
    Mohammadian, Amin
    Aghaeinia, Hassan
    Towhidkhah, Farzad
    IET IMAGE PROCESSING, 2015, 9 (07) : 596 - 603
  • [8] Video-Based Facial Expression Recognition Using Hough Forest
    Hsu, Shih-Chung
    Hsu, Chi-Ting
    Huang, Chung-Lin
    JOURNAL OF INFORMATION SCIENCE AND ENGINEERING, 2016, 32 (02) : 477 - 494
  • [9] Automated facial video-based recognition of depression and anxiety symptom severity: cross-corpus validation
    A. Pampouchidou
    M. Pediaditis
    E. Kazantzaki
    S. Sfakianakis
    I. A. Apostolaki
    K. Argyraki
    D. Manousos
    F. Meriaudeau
    K. Marias
    F. Yang
    M. Tsiknakis
    M. Basta
    A. N. Vgontzas
    P. Simos
    Machine Vision and Applications, 2020, 31
  • [10] Automated facial video-based recognition of depression and anxiety symptom severity: cross-corpus validation
    Pampouchidou, A.
    Pediaditis, M.
    Kazantzaki, E.
    Sfakianakis, S.
    Apostolaki, I. A.
    Argyraki, K.
    Manousos, D.
    Meriaudeau, F.
    Marias, K.
    Yang, F.
    Tsiknakis, M.
    Basta, M.
    Vgontzas, A. N.
    Simos, P.
    MACHINE VISION AND APPLICATIONS, 2020, 31 (04)