Loss Relaxation Strategy for Noisy Facial Video-based Automatic Depression Recognition

被引:0
|
作者
Song S. [1 ]
Luo Y. [2 ]
Tumer T. [3 ]
Fu C. [4 ]
Valstar M. [5 ]
Gunes H. [6 ]
机构
[1] University of Cambridge, Cambridge, Leicester
[2] Imperial College London, London
[3] Middle East Technical University, Ankara
[4] Northeastern University, Qinhuangdao
[5] University of Nottingham, Nottingham
[6] University of Cambridge, Cambridge
来源
基金
英国工程与自然科学研究理事会;
关键词
Depression classification; depression severity estimation; facial video analysis; loss relaxation strategy; noisy data and annotation;
D O I
10.1145/3648696
中图分类号
学科分类号
摘要
Automatic depression analysis has been widely investigated on face videos that have been carefully collected and annotated in lab conditions. However, videos collected under real-world conditions may suffer from various types of noise due to challenging data acquisition conditions and lack of annotators. Although deep learning (DL) models frequently show excellent depression analysis performances on datasets collected in controlled lab conditions, such noise may degrade their generalization abilities for real-world depression analysis tasks. In this article, we uncovered that noisy facial data and annotations consistently change the distribution of training losses for facial depression DL models; i.e., noisy data-label pairs cause larger loss values compared to clean data-label pairs. Since different loss functions could be applied depending on the employed model and task, we propose a generic loss function relaxation strategy that can jointly reduce the negative impact of various noisy data and annotation problems occurring in both classification and regression loss functions for face video-based depression analysis, where the parameters of the proposed strategy can be automatically adapted during depression model training. The experimental results on 25 different artificially created noisy depression conditions (i.e., five noise types with five different noise levels) show that our loss relaxation strategy can clearly enhance both classification and regression loss functions, enabling the generation of superior face video-based depression analysis models under almost all noisy conditions. Our approach is robust to its main variable settings and can adaptively and automatically obtain its parameters during training. © 2024 Copyright held by the owner/author(s).
引用
收藏
相关论文
共 50 条
  • [21] Video-Based Facial Weakness Analysis
    Zhuang, Yan
    McDonald, Mark M.
    Aldridge, Chad M.
    Hassan, Mohamed Abul
    Uribe, Omar
    Arteaga, Daniel
    Southerland, Andrew M.
    Rohde, Gustavo K.
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2021, 68 (09) : 2698 - 2705
  • [22] A Deep Spatial and Temporal Aggregation Framework for Video-Based Facial Expression Recognition
    Pan, Xianzhang
    Ying, Guoliang
    Chen, Guodong
    Li, Hongming
    Li, Wenshu
    IEEE ACCESS, 2019, 7 : 48807 - 48815
  • [23] Deep Temporal-Spatial Aggregation for Video-Based Facial Expression Recognition
    Pan, Xianzhang
    Guo, Wenping
    Guo, Xiaoying
    Li, Wenshu
    Xu, Junjie
    Wu, Jinzhao
    SYMMETRY-BASEL, 2019, 11 (01):
  • [24] Cross-Dimension Transfer Learning for Video-Based Facial Expression Recognition
    Zhong, Kailun
    Li, Yi
    Fang, Li
    Chen, Ping
    BIOMETRIC RECOGNITION (CCBR 2019), 2019, 11818 : 180 - 189
  • [25] STATISTICAL, SPECTRAL AND GRAPH REPRESENTATIONS FOR VIDEO-BASED FACIAL EXPRESSION RECOGNITION IN CHILDREN
    Abbasi, Nida Itrat
    Song, Siyang
    Gunes, Hatice
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 1725 - 1729
  • [26] Learning Spatiotemporal and Geometric Features with ISA for Video-Based Facial Expression Recognition
    Lin, Chenhan
    Long, Fei
    Yao, Junfeng
    Sun, Ming-Ting
    Su, Jinsong
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT III, 2017, 10636 : 435 - 444
  • [27] A Unified Framework for Monocular Video-Based Facial Motion Tracking and Expression Recognition
    Yu, Jun
    MULTIMEDIA MODELING, MMM 2017, PT II, 2017, 10133 : 50 - 62
  • [28] Video-based framework for face recognition in video
    Gorodnichy, DO
    2ND CANADIAN CONFERENCE ON COMPUTER AND ROBOT VISION, PROCEEDINGS, 2005, : 330 - 338
  • [29] Analysis of facial expressions in parkinson's disease through video-based automatic methods
    Bandini, Andrea
    Orlandi, Silvia
    Escalante, Hugo Jair
    Giovannelli, Fabio
    Cincotta, Massimo
    Reyes-Garcia, Carlos A.
    Vanni, Paola
    Zaccara, Gaetano
    Manfredi, Claudia
    JOURNAL OF NEUROSCIENCE METHODS, 2017, 281 : 7 - 20
  • [30] Adaptive metric learning with deep neural networks for video-based facial expression recognition
    Liu, Xiaofeng
    Ge, Yubin
    Yang, Chao
    Jia, Ping
    JOURNAL OF ELECTRONIC IMAGING, 2018, 27 (01)