Research on Wearable Emotion Recognition Based on Multi-Source Domain Adversarial Transfer Learning

被引:0
|
作者
Zou Y.-P. [1 ]
Wang D.-Y. [1 ]
Wang D. [1 ]
Zheng C.-L. [1 ]
Song Q.-F. [1 ]
Zhu Y.-Z. [1 ]
Fan C.-H. [2 ]
Wu K.-S. [3 ]
机构
[1] IoT Research Center, College of Computer Science and Software Engineering, ShenzhenUniversity, Guangdong, Shenzhen
[2] Department of Psychiatry, Guangdong Second Provincial General Hospital, Guangzhou
[3] Information Hub, Hong Kong University of Science and Technology(guangzhou), Guangzhou
来源
基金
中国国家自然科学基金;
关键词
domain adaptation; emotion recognition; generative-adversarial learning; multimodal data; transfer learning; wearable devices;
D O I
10.11897/SP.J.1016.2024.00266
中图分类号
学科分类号
摘要
Emotions can profoundly impact both human’s overall well-being and cognitive function. As a result, they are of paramount significance in the realm of human life especially in modern society with increasing pressures. Automatic emotion recognition contributes to early warning of psychological disorders and the exploration of behavioral mechanisms, holding immense research and practical value. Over the past decade, researchers have proposed various kinds of methods for automatic emotion recognition based on different sensing mechanisms. Nevertheless, each of them exhibits deficiencies in different aspects. For example, the methods based on electroencephalogram (EEG) signals require the use of specialized, costly, and challenging-to-operate EEG devices;the methods relying on visual and speech cues carry privacy risks;and the methods based on the analysis of mobile phone usage pattern need improvement in terms of reliability and accuracy. Considering the above,this paper proposes a novel approach to automatic emotion recognition that utilizes low-cost,readily available,and easy-to-use wearable hardware. In a detail, this paper makes use of the potential correlations between physiological signals,namely,breathing and heartbeat sounds,and pulse with human emotions. By employing data fusion across multiple sensing modalities,this work effectively harnesses diverse information types,reducing data redundancy,and substantially improving the system performance at the same time. Furthermore, while ensuring a high recognition accuracy, this paper also proposes an emotion recognition model based on a multi-source domain adversarial approach which aims to enhance the generalization of emotion recognition across diverse users and minimize the cost for unseen users. Our method first leverages a small amount of unlabeled data from unseen users to achieve quick model adaptation in an unsupervised approach, and then fine-tune the classifier’s parameters with a minimal amount of labeled data to further improve emotion recognition accuracy. To validate the effectiveness of our proposed emotion recognition method, this paper designs and implements a wearable system that integrate two microphones and photoplethysmography(PPG)sensors to measure physiological signs. Among them, the two microphones are equipped in a smartglasses and earphone to collect sounds produced by heartbeats and breathing, respectively;the two PPG sensors are embedded in the smartglasses and a smartwatch to measure the blood pulses in the head and wrist, respectively. Based on this wearable system, we have conducted extensive experiments in diverse settings with thirty participants aged from 17 to 30 years old. We have also carried an assessment of the impact of different environmental factors such as noise, hardware, and wearing positions to evaluate the robustness of our emotion recognition system. The experimental results demonstrate that for the four basic emotions, the proposed method achieves an average recognition accuracy of 95. 0%in the subject-dependent cases, and an average accuracy of 62. 5% in the cross-subject cases after using multi-source domain adversarial transfer learning,with a 5. 3% improvement over the baseline methods. When combined with supervised fine-tuning with few shots, the recognition accuracy further increases to 81. 1%,surpassing the baseline methods by 12. 4%. These findings affirm the feasibility of the proposed method and offer a fresh perspective for ubiquitous emotion recognition research. © 2024 Science Press. All rights reserved.
引用
收藏
页码:266 / 286
页数:20
相关论文
共 69 条
  • [1] Xue-Feng Chen, Xiao-Lan Fu, Kan Zhang, Report on National Mental Health Development In China(2019-2020), (2021)
  • [2] Plutchik R., Emotions and Life:Perspectives from Psychology,Biology, and Evolution, (2003)
  • [3] Russell J A., A circumplex model of affect, Journal of Personality and Social Psychology, 39, 6, pp. 1161-1178, (1980)
  • [4] Mehendale N., Facial emotion recognition using convolutional neural networks(FERC), SN Applied Sciences, 2, 3, pp. 1-8, (2020)
  • [5] Hickson S, ,Kwatra V,Dufour N,Sud A,Essa I. Eyemotion:classifying facial expressions in VR using eye-tracking cameras, Proceedings of the IEEE Winter Conference on Applications of Computer Vision, pp. 1626-1635, (2019)
  • [6] EMO:real-time emotion recognition from single-eye images for resource-constrained eyewear devices, Proceedings of the ACM International Conference on Mobile Systems,Applications,and Services, pp. 448-461, (2020)
  • [7] A study on emotion recognition from body gestures using Kinect sensor, Proceedings of the International Conference on Communication and Signal Processing, pp. 56-60, (2014)
  • [8] Castellano G,, Villalba S D,, Camurri A., Recognising human emotions from body movement and gesture dynamics, Proceedings of the International Conference on Affective Computing and Intelligent Interaction, pp. 71-82, (2007)
  • [9] Montepare J M,, Goldstein S B,, Clausen A., The identification of emotions from gait information, Journal of Nonverbal Behavior, 11, 1, pp. 33-42, (1987)
  • [10] Lalitha S,, Geyasruti D,, Narayanan R,, Et al., Emotion detection using MFCC and cepstrum features, Procedia Computer Science, 70, pp. 29-35, (2015)