Graph to Grid: Learning Deep Representations for Multimodal Emotion Recognition

被引:1
|
作者
Jin, Ming [1 ]
Li, Jinpeng [1 ]
机构
[1] Univ Chinese Acad Sci, Ningbo Inst Life & Hlth Ind, Ningbo Hosp 2, Ningbo, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Brain-computer Interface; Electroencephalogram; Multimodal Emotion Recognition;
D O I
10.1145/3581783.3612074
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal emotion recognition based on electroencephalogram (EEG) and compensating physiological signals (e.g., eye tracking) has shown potential in the diagnosis and rehabilitation tracking of depression. Since the multi-channel EEG signals are generally processed as one-dimensional (1-D) graph-like features, existing approaches can only adopt underdeveloped shallow models to recognize emotions. However, these simple models have difficulty decoupling complex emotion patterns due to their limited representation capacity. To address this problem, we propose the graph-to-grid (G2G), a concise and plug-and-play module that transforms the 1-D graph-like data into the two-dimensional (2-D) grid-like data via the numerical relation coding. After that, the well developed deep models, e.g., ResNet can be used to downstream tasks. In addition, G2G simplifies the previous complex multimodal fusion into an input matrix augmentation operation, which greatly reduces the difficulty of model design and parameter tuning. Extensive results on three public datasets (SEED, SEED5 and MPED) indicate that the proposed approach achieves state-of-the-art emotion recognition accuracy in both unimodal and multimodal settings, with good cross-session generalization ability. G2G enables the development of more appropriate multimodal emotion recognition algorithms for follow-up studies. Our code is publicly available at https://github.com/Jinminbox/G2G.
引用
收藏
页码:5985 / 5993
页数:9
相关论文
共 50 条
  • [1] Emotion Recognition Using Multimodal Deep Learning
    Liu, Wei
    Zheng, Wei-Long
    Lu, Bao-Liang
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2016, PT II, 2016, 9948 : 521 - 529
  • [2] Emotion Recognition on Multimodal with Deep Learning and Ensemble
    Dharma, David Adi
    Zahra, Amalia
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (12) : 656 - 663
  • [3] Emotion Recognition on Multimodal with Deep Learning and Ensemble
    Dharma, David Adi
    Zahra, Amalia
    [J]. International Journal of Advanced Computer Science and Applications, 2022, 13 (12): : 656 - 663
  • [4] DEEP MULTIMODAL LEARNING FOR EMOTION RECOGNITION IN SPOKEN LANGUAGE
    Gu, Yue
    Chen, Shuhong
    Marsic, Ivan
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5079 - 5083
  • [5] Multimodal Arabic emotion recognition using deep learning
    Al Roken, Noora
    Barlas, Gerassimos
    [J]. SPEECH COMMUNICATION, 2023, 155
  • [6] Multimodal Emotion Recognition using Deep Learning Architectures
    Ranganathan, Hiranmayi
    Chakraborty, Shayok
    Panchanathan, Sethuraman
    [J]. 2016 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2016), 2016,
  • [7] Annotation Efficiency in Multimodal Emotion Recognition with Deep Learning
    Zhu, Lili
    Spachos, Petros
    [J]. 2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 560 - 565
  • [8] Comparing Recognition Performance and Robustness of Multimodal Deep Learning Models for Multimodal Emotion Recognition
    Liu, Wei
    Qiu, Jie-Lin
    Zheng, Wei-Long
    Lu, Bao-Liang
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (02) : 715 - 729
  • [9] EmoNets: Multimodal deep learning approaches for emotion recognition in video
    Kahou, Samira Ebrahimi
    Bouthillier, Xavier
    Lamblin, Pascal
    Gulcehre, Caglar
    Michalski, Vincent
    Konda, Kishore
    Jean, Sebastien
    Froumenty, Pierre
    Dauphin, Yann
    Boulanger-Lewandowski, Nicolas
    Ferrari, Raul Chandias
    Mirza, Mehdi
    Warde-Farley, David
    Courville, Aaron
    Vincent, Pascal
    Memisevic, Roland
    Pal, Christopher
    Bengio, Yoshua
    [J]. JOURNAL ON MULTIMODAL USER INTERFACES, 2016, 10 (02) : 99 - 111
  • [10] EmoNets: Multimodal deep learning approaches for emotion recognition in video
    Samira Ebrahimi Kahou
    Xavier Bouthillier
    Pascal Lamblin
    Caglar Gulcehre
    Vincent Michalski
    Kishore Konda
    Sébastien Jean
    Pierre Froumenty
    Yann Dauphin
    Nicolas Boulanger-Lewandowski
    Raul Chandias Ferrari
    Mehdi Mirza
    David Warde-Farley
    Aaron Courville
    Pascal Vincent
    Roland Memisevic
    Christopher Pal
    Yoshua Bengio
    [J]. Journal on Multimodal User Interfaces, 2016, 10 : 99 - 111