Soft Voting Strategy for Multi-Modal Emotion Recognition Using Deep-learning- Facial Images and EEG

被引:1
|
作者
Chinta, Uma [1 ]
Kalita, Jugal [1 ]
Atyabi, Adham [1 ]
机构
[1] Univ Colorado, Dept Comp Sci, Colorado Springs, CO 80907 USA
关键词
EEG; feature extraction; emotion analysis; multi-modal integration; Gated Recurrent Unit; FUSION;
D O I
10.1109/CCWC57344.2023.10099070
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion recognition is an important factor in social communication and has a wide range of applications from retail to healthcare. In psychology, emotion recognition focuses on emotional states within non-verbal visual and auditory cues. It is essential to the human ability to associate meaning with events rather than treating them as mere facts. Studies of emotion recognition often utilize data gathered in response to non-verbal cues using modalities such as eye tracking, Electroencephalogram (EEG), and facial video and build classification models capable of differentiating responses to various emotions and cues. The accuracy of these emotion recognition models largely depends on feature representation and the suitability of the chosen features in magnifying the differences between patterns of various emotions. Single-modal feature extraction methods are limited in capturing between-group differences and often result in reduced classification performance. To address this problem, this paper proposes a multi-modal approach in the representation of response to emotional cues involving EEG recording and facial video data. The study utilizes the dataset containing frontal face video recordings and EEG data of 22 participants. A novel deep neural network architecture within the feature level is used to efficiently predict emotions using EEG and facial video data. The experimental result indicates 97.5% accuracy in identifying facial expressions and categorizing them into two classes, arousal (class 0) and valence (class 1), surpassing state-of-the-art for the DEAP dataset.
引用
收藏
页码:738 / 745
页数:8
相关论文
共 50 条
  • [1] Multi-Modal Emotion Recognition Based On deep Learning Of EEG And Audio Signals
    Li, Zhongjie
    Zhang, Gaoyan
    Dang, Jianwu
    Wang, Longbiao
    Wei, Jianguo
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [2] Facial emotion recognition using multi-modal information
    De Silva, LC
    Miyasato, T
    Nakatsu, R
    ICICS - PROCEEDINGS OF 1997 INTERNATIONAL CONFERENCE ON INFORMATION, COMMUNICATIONS AND SIGNAL PROCESSING, VOLS 1-3: THEME: TRENDS IN INFORMATION SYSTEMS ENGINEERING AND WIRELESS MULTIMEDIA COMMUNICATIONS, 1997, : 397 - 401
  • [3] A Multi-Modal Deep Learning Approach for Emotion Recognition
    Shahzad, H. M.
    Bhatti, Sohail Masood
    Jaffar, Arfan
    Rashid, Muhammad
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 36 (02): : 1561 - 1570
  • [4] Multi-Modal Emotion Recognition From Speech and Facial Expression Based on Deep Learning
    Cai, Linqin
    Dong, Jiangong
    Wei, Min
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 5726 - 5729
  • [5] Multi-modal emotion recognition using EEG and speech signals
    Wang, Qian
    Wang, Mou
    Yang, Yan
    Zhang, Xiaolei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 149
  • [6] A multi-modal deep learning system for Arabic emotion recognition
    Abu Shaqra F.
    Duwairi R.
    Al-Ayyoub M.
    International Journal of Speech Technology, 2023, 26 (01) : 123 - 139
  • [7] Multi-modal emotion identification fusing facial expression and EEG
    Wu, Yongzhen
    Li, Jinhua
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (07) : 10901 - 10919
  • [8] Multi-modal emotion identification fusing facial expression and EEG
    Yongzhen Wu
    Jinhua Li
    Multimedia Tools and Applications, 2023, 82 : 10901 - 10919
  • [9] Intelligent ear for emotion recognition: Multi-modal emotion recognition via acoustic features, semantic contents and facial images
    Wu, CH
    Chuang, ZJ
    8TH WORLD MULTI-CONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL XII, PROCEEDINGS: APPLICATIONS OF CYBERNETICS AND INFORMATICS IN OPTICS, SIGNALS, SCIENCE AND ENGINEERING, 2004, : 122 - 127
  • [10] Multi-modal embeddings using multi-task learning for emotion recognition
    Khare, Aparna
    Parthasarathy, Srinivas
    Sundaram, Shiva
    INTERSPEECH 2020, 2020, : 384 - 388