Multimodal Depression Detection Based on Self-Attention Network With Facial Expression and Pupil

被引:0
|
作者
Liu, Xiang [1 ]
Shen, Hao [2 ]
Li, Huiru [2 ]
Tao, Yongfeng [2 ]
Yang, Minqiang [2 ]
机构
[1] Dongguan Univ Technol, Sch Comp Sci & Technol, Dongguan 523000, Peoples R China
[2] Lanzhou Univ, Sch Informat Sci & Engn, Lanzhou 730000, Peoples R China
基金
中国国家自然科学基金;
关键词
Depression; Feature extraction; Pupils; Face recognition; Accuracy; Electroencephalography; Data mining; Facial expression; multimodal fusion; pupil; self-attention network; CANONICAL CORRELATION-ANALYSIS; RECOGNITION; FUSION;
D O I
10.1109/TCSS.2024.3405949
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Depression is a major mental health issue in contemporary society, with an estimated 350 million people affected globally. The number of individuals diagnosed with depression continues to rise each year. Currently, clinical practice relies entirely on self-reporting and clinical assessment, which carries the risk of subjective biases. In this article, we propose a multimodal method based on facial expression and pupil to detect depression more objectively and precisely. Our method first extracts the features of facial expressions and pupil diameter using residual networks and 1-D convolutional neural networks. Second, a cross-modal fusion model based on self-attention networks (CMF-SNs) is proposed, which utilizes cross-modal attention networks within modalities and parallel self-attention networks between different modalities to extract CMF features of facial expressions and pupil diameter, effectively complementing information between different modalities. Finally, the obtained features are fully connected to identify depression. Multiple controlled experiments show that compared to single modality, the multimodal fusion method based on self-attention networks has a higher ability to recognize depression, with the highest accuracy of 75.0%. In addition, we conducted comparative experiments under three different stimulation paradigms, and the results showed that the classification accuracy under negative and neutral stimuli was higher than that under positive stimuli, indicating a bias of depressed patients toward negative images. The experimental results demonstrate the superiority of our multimodal fusion method.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] A visual self-attention network for facial expression recognition
    Yu, Naigong
    Bai, Deguo
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [2] A framework for facial expression recognition using deep self-attention network
    Indolia S.
    Nigam S.
    Singh R.
    [J]. Journal of Ambient Intelligence and Humanized Computing, 2023, 14 (07): : 9543 - 9562
  • [3] Regional Self-Attention Convolutional Neural Network for Facial Expression Recognition
    Zhou, Lifang
    Wang, Yi
    Lei, Bangjun
    Yang, Weibin
    [J]. INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (08)
  • [4] A self-attention network for smoke detection
    Jiang, Minghua
    Zhao, Yaxin
    Yu, Feng
    Zhou, Changlong
    Peng, Tao
    [J]. FIRE SAFETY JOURNAL, 2022, 129
  • [5] Multimodal cooperative self-attention network for action recognition
    Zhong, Zhuokun
    Hou, Zhenjie
    Liang, Jiuzhen
    Lin, En
    Shi, Haiyong
    [J]. IET IMAGE PROCESSING, 2023, 17 (06) : 1775 - 1783
  • [6] Network Intrusion Detection Based on Self-Attention Mechanism and BIGRU
    Du, Xuran
    Gan, Gang
    [J]. 2024 2ND INTERNATIONAL CONFERENCE ON MOBILE INTERNET, CLOUD COMPUTING AND INFORMATION SECURITY, MICCIS 2024, 2024, : 236 - 241
  • [7] Research on Anomaly Network Detection Based on Self-Attention Mechanism
    Hu, Wanting
    Cao, Lu
    Ruan, Qunsheng
    Wu, Qingfeng
    [J]. SENSORS, 2023, 23 (11)
  • [8] Self-attention recurrent network for saliency detection
    Sun, Fengdong
    Li, Wenhui
    Guan, Yuanyuan
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (21) : 30793 - 30807
  • [9] Self-attention recurrent network for saliency detection
    Fengdong Sun
    Wenhui Li
    Yuanyuan Guan
    [J]. Multimedia Tools and Applications, 2019, 78 : 30793 - 30807
  • [10] Multimodal Fusion Method Based on Self-Attention Mechanism
    Zhu, Hu
    Wang, Ze
    Shi, Yu
    Hua, Yingying
    Xu, Guoxia
    Deng, Lizhen
    [J]. WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2020, 2020