Multi-modal emotion recognition using EEG and speech signals

被引:28
|
作者
Wang, Qian [1 ]
Wang, Mou [1 ]
Yang, Yan [1 ]
Zhang, Xiaolei [1 ]
机构
[1] Northwestern Polytech Univ, Xian 710072, Shaanxi, Peoples R China
关键词
Multi-modal emotion database; EEG emotion recognition; Speech emotion recognition; Physiological signal; Data fusion; IDENTIFICATION; DATABASE; MACHINE;
D O I
10.1016/j.compbiomed.2022.105907
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Automatic Emotion Recognition (AER) is critical for naturalistic Human-Machine Interactions (HMI). Emotions can be detected through both external behaviors, e.g., tone of voice and internal physiological signals, e.g., electroencephalogram (EEG). In this paper, we first constructed a multi-modal emotion database, named Multi -modal Emotion Database with four modalities (MED4). MED4 consists of synchronously recorded signals of participants' EEG, photoplethysmography, speech and facial images when they were influenced by video stimuli designed to induce happy, sad, angry and neutral emotions. The experiment was performed with 32 participants in two environment conditions, a research lab with natural noises and an anechoic chamber. Four baseline algorithms were developed to verify the database and the performances of AER methods, Identification-vector + Probabilistic Linear Discriminant Analysis (I-vector + PLDA), Temporal Convolutional Network (TCN), Extreme Learning Machine (ELM) and Multi-Layer Perception Network (MLP). Furthermore, two fusion strategies on feature-level and decision-level respectively were designed to utilize both external and internal information of human status. The results showed that EEG signals generate higher accuracy in emotion recognition than that of speech signals (achieving 88.92% in anechoic room and 89.70% in natural noisy room vs 64.67% and 58.92% respectively). Fusion strategies that combine speech and EEG signals can improve overall accuracy of emotion recognition by 25.92% when compared to speech and 1.67% when compared to EEG in anechoic room and 31.74% and 0.96% in natural noisy room. Fusion methods also enhance the robustness of AER in the noisy environment. The MED4 database will be made publicly available, in order to encourage researchers all over the world to develop and validate various advanced methods for AER.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Multi-modal emotion identification fusing facial expression and EEG
    Yongzhen Wu
    Jinhua Li
    [J]. Multimedia Tools and Applications, 2023, 82 : 10901 - 10919
  • [42] Multi-modal emotion identification fusing facial expression and EEG
    Wu, Yongzhen
    Li, Jinhua
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (07) : 10901 - 10919
  • [43] Lightweight multi-modal emotion recognition model based on modal generation
    Liu, Peisong
    Che, Manqiang
    Luo, Jiangchuan
    [J]. 2022 9TH INTERNATIONAL FORUM ON ELECTRICAL ENGINEERING AND AUTOMATION, IFEEA, 2022, : 430 - 435
  • [44] Cross-modal dynamic convolution for multi-modal emotion recognition
    Wen, Huanglu
    You, Shaodi
    Fu, Ying
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 78
  • [45] EEG-Based Multi-Modal Emotion Recognition using Bag of Deep Features: An Optimal Feature Selection Approach
    Asghar, Muhammad Adeel
    Khan, Muhammad Jamil
    Fawad
    Amin, Yasar
    Rizwan, Muhammad
    Rahman, MuhibUr
    Badnava, Salman
    Mirjavadi, Seyed Sajad
    [J]. SENSORS, 2019, 19 (23)
  • [46] Audio-Visual Emotion Recognition System Using Multi-Modal Features
    Handa, Anand
    Agarwal, Rashi
    Kohli, Narendra
    [J]. INTERNATIONAL JOURNAL OF COGNITIVE INFORMATICS AND NATURAL INTELLIGENCE, 2021, 15 (04)
  • [47] HUMAN EMOTION RECOGNITION USING MULTI-MODAL BIOLOGICAL SIGNALS BASED ON TIME LAG-CONSIDERED CORRELATION MAXIMIZATION
    Moroto, Yuya
    Maeda, Keisuke
    Ogawa, Takahiro
    Haseyama, Miki
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4683 - 4687
  • [48] Interpretable Emotion Recognition Using EEG Signals
    Qing, Chunmei
    Qiao, Rui
    Xu, Xiangmin
    Cheng, Yongqiang
    [J]. IEEE ACCESS, 2019, 7 : 94160 - 94170
  • [49] Multi-Modal Domain Adaptation Variational Auto-encoder for EEG-Based Emotion Recognition
    Wang, Yixin
    Qiu, Shuang
    Li, Dan
    Du, Changde
    Lu, Bao-Liang
    He, Huiguang
    [J]. IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2022, 9 (09) : 1612 - 1626
  • [50] Multi-modal speech emotion detection using optimised deep neural network classifier
    Padman, Sweta Nishant
    Magare, Dhiraj
    [J]. COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING-IMAGING AND VISUALIZATION, 2023, : 2020 - 2038