Semi-Supervised Cross-Subject Emotion Recognition Based on Stacked Denoising Autoencoder Architecture Using a Fusion of Multi-Modal Physiological Signals

被引:10
|
作者
Luo, Junhai [1 ]
Tian, Yuxin [1 ]
Yu, Hang [1 ]
Chen, Yu [1 ]
Wu, Man [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 610056, Peoples R China
关键词
DEAP dataset; electroencephalogram (EEG); emotion recognition; multi-source fusion; stacked denoising autoencoder; unsupervised representation learning; FEATURE-EXTRACTION; TIME-SERIES; EEG; REPRESENTATIONS;
D O I
10.3390/e24050577
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
In recent decades, emotion recognition has received considerable attention. As more enthusiasm has shifted to the physiological pattern, a wide range of elaborate physiological emotion data features come up and are combined with various classifying models to detect one's emotional states. To circumvent the labor of artificially designing features, we propose to acquire affective and robust representations automatically through the Stacked Denoising Autoencoder (SDA) architecture with unsupervised pre-training, followed by supervised fine-tuning. In this paper, we compare the performances of different features and models through three binary classification tasks based on the Valence-Arousal-Dominance (VAD) affection model. Decision fusion and feature fusion of electroencephalogram (EEG) and peripheral signals are performed on hand-engineered features; data-level fusion is performed on deep-learning methods. It turns out that the fusion data perform better than the two modalities. To take advantage of deep-learning algorithms, we augment the original data and feed it directly into our training model. We use two deep architectures and another generative stacked semi-supervised architecture as references for comparison to test the method's practical effects. The results reveal that our scheme slightly outperforms the other three deep feature extractors and surpasses the state-of-the-art of hand-engineered features.
引用
收藏
页数:29
相关论文
共 26 条
  • [21] Multi-modal Fusion Methods for Robust Emotion Recognition using Body-worn Physiological Sensors in Mobile Environments
    Zhang, Tianyi
    [J]. ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, : 463 - 467
  • [22] A Parallel Multi-Modal Factorized Bilinear Pooling Fusion Method Based on the Semi-Tensor Product for Emotion Recognition
    Liu, Fen
    Chen, Jianfeng
    Li, Kemeng
    Tan, Weijie
    Cai, Chang
    Ayub, Muhammad Saad
    [J]. ENTROPY, 2022, 24 (12)
  • [23] Joint low-rank tensor fusion and cross-modal attention for multimodal physiological signals based emotion recognition
    Wan, Xin
    Wang, Yongxiong
    Wang, Zhe
    Tang, Yiheng
    Liu, Benke
    [J]. PHYSIOLOGICAL MEASUREMENT, 2024, 45 (07)
  • [24] HUMAN EMOTION RECOGNITION USING MULTI-MODAL BIOLOGICAL SIGNALS BASED ON TIME LAG-CONSIDERED CORRELATION MAXIMIZATION
    Moroto, Yuya
    Maeda, Keisuke
    Ogawa, Takahiro
    Haseyama, Miki
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4683 - 4687
  • [25] New semi-supervised classification using a multi-modal feature joint L21-norm based sparse representation
    Cui, Yan
    Jiang, Jielin
    Lai, Zhihui
    Hu, Zuojin
    Jiang, Yuquan
    Wong, WaiKeung
    [J]. SIGNAL PROCESSING-IMAGE COMMUNICATION, 2018, 65 : 94 - 106
  • [26] Deep time-frequency features and semi-supervised dimension reduction for subject-independent emotion recognition from multi-channel EEG signals
    Zali-Vargahan, Behrooz
    Charmin, Asghar
    Kalbkhani, Hashem
    Barghandan, Saeed
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 85