Recognition of emotions using multimodal physiological signals and an ensemble deep learning model

被引:253
|
作者
Yin, Zhong [1 ]
Zhao, Mengyuan [2 ]
Wang, Yongxiong [1 ]
Yang, Jingdong [1 ]
Zhang, Jianhua [3 ]
机构
[1] Univ Shanghai Sci & Technol, Minist Educ, Engn Res Ctr Opt Instrument & Syst, Shanghai Key Lab Modern Opt Syst, Shanghai 200093, Peoples R China
[2] Univ Shanghai Sci & Technol, Sch Social Sci, Shanghai 200093, PR, Peoples R China
[3] East China Univ Sci & Technol, Dept Automat, Shanghai 200237, PR, Peoples R China
基金
中国国家自然科学基金;
关键词
Emotion recognition; Affective computing; Physiological signals; Deep learning; Ensemble learning; EEG; CLASSIFICATION; VECTOR; BCI; FUSION;
D O I
10.1016/j.cmpb.2016.12.005
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Background and Objective: Using deep-learning methodologies to analyze multimodal physiological signals becomes increasingly attractive for recognizing human emotions. However, the conventional deep emotion classifiers may suffer from the drawback of the lack of the expertise for determining model structure and the oversimplification of combining multimodal feature abstractions. Methods: In this study, a multiple-fusion-layer based ensemble classifier of stacked autoencoder (MESAE) is proposed for recognizing emotions, in which the deep structure is identified based on a physiological data-driven approach. Each SAE consists of three hidden layers to filter the unwanted noise in the physiological features and derives the stable feature representations. An additional deep model is used to achieve the SAE ensembles. The physiological features are split into several subsets according to different feature extraction approaches with each subset separately encoded by a SAE. The derived SAE abstractions are combined according to the physiological modality to create six sets of encodings, which are then fed to a three-layer, adjacent-graph-based network for feature fusion. The fused features are used to recognize binary arousal or valence states. Results: DEAP multimodal database was employed to validate the performance of the MESAE. By comparing with the best existing emotion classifier, the mean of classification rate and F-score improves by 5.26%. Conclusions: The superiority of the MESAE against the state-of-the-art shallow and deep emotion classifiers has been demonstrated under different sizes of the available physiological instances. (C) 2016 Elsevier Ireland Ltd. All rights reserved.
引用
收藏
页码:93 / 110
页数:18
相关论文
共 50 条
  • [21] Emotion Recognition Using Multimodal Deep Learning
    Liu, Wei
    Zheng, Wei-Long
    Lu, Bao-Liang
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2016, PT II, 2016, 9948 : 521 - 529
  • [22] Multimodal Deep Feature Aggregation for Facial Action Unit Recognition using Visible Images and Physiological Signals
    Lakshminarayana, Nagashri N.
    Sankaran, Nishant
    Setlur, Srirangaraj
    Govindaraju, Venu
    [J]. 2019 14TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2019), 2019, : 458 - 461
  • [23] Ensemble deep learning model for optical character recognition
    Shetty, Ashish
    Sharma, Sanjeev
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 11411 - 11431
  • [24] Quaternary classification of emotions based on electroencephalogram signals using hybrid deep learning model
    Singh K.
    Ahirwal M.K.
    Pandey M.
    [J]. Journal of Ambient Intelligence and Humanized Computing, 2023, 14 (03) : 2429 - 2441
  • [25] Multimodal Emotion Recognition Using Visual, Vocal and Physiological Signals: A Review
    Udahemuka, Gustave
    Djouani, Karim
    Kurien, Anish M.
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (17):
  • [26] Mathematical representation of emotion using multimodal recognition model with deep multitask learning
    モダリティを統合した認識モデルに基づく深層マルチタスク学習による感情の数理的表現
    [J]. Harata, Seiichi (harata@katolab.nitech.ac.jp), 1600, Institute of Electrical Engineers of Japan (140): : 1343 - 1351
  • [27] Ensemble deep learning model for optical character recognition
    Ashish Shetty
    Sanjeev Sharma
    [J]. Multimedia Tools and Applications, 2024, 83 : 11411 - 11431
  • [28] Hierarchical extreme puzzle learning machine-based emotion recognition using multimodal physiological signals
    Pradhan, Anushka
    Srivastava, Subodh
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 83
  • [29] Multimodal Fake-News Recognition Using Ensemble of Deep Learners
    Al Obaid, Abdulhameed
    Khotanlou, Hassan
    Mansoorizadeh, Muharram
    Zabihzadeh, Davood
    [J]. ENTROPY, 2022, 24 (09)
  • [30] Multimodal Recognition of Emotions with Application to Mobile Learning
    Lucia Barron-Estrada, Maria
    Zatarain-Cabada, Ramon
    Aispuro-Gallegos, Claudia G.
    [J]. 2018 IEEE 18TH INTERNATIONAL CONFERENCE ON ADVANCED LEARNING TECHNOLOGIES (ICALT 2018), 2018, : 416 - 418