Incongruity-aware multimodal physiology signals fusion for emotion recognition

被引:0
|
作者
Li, Jing [1 ]
Chen, Ning [1 ]
Zhu, Hongqing [1 ]
Li, Guangqiang [1 ]
Xu, Zhangyong [1 ]
Chen, Dingxin [1 ]
机构
[1] East China Univ Sci & Technol, Sch Informat Sci & Engn, Shanghai 200237, Peoples R China
关键词
Emotion recognition; Multi-modal fusion; Physiological signal incongruity; Cross-Modal Transformer (CMT); Self-Attention Transformer (SAT); Low Rank Fusion (LRF); EEG;
D O I
10.1016/j.inffus.2023.102220
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Various physiological signals can reflect the human's emotional states objectively. How to take advantage of the common as well as complementary properties of different physiological signals in representing the emotional states is an interesting problem. Although various models have been constructed to fuse multimodal physiological signals for emotion recognition, the possible incongruity existing among different physiological signals in representing the emotional states and the redundancy resulted from the fusion, which may affect the performance of the fusion schemes seriously, were seldom considered. To this end, a fusion model, which can eliminate the incongruity among different physiological signals and reduce the information redundancy to some extent, is proposed. First, one physiological signal is chosen as the primary modality due to its prominent performance in emotion recognition, and the remaining physiological signals are viewed as the auxiliary modalities. Secondly, the Cross Modal Transformer (CMT) is adopted to optimize the features of the auxiliary modalities by eliminating the incongruity among them, and then Low Rank Fusion (LRF) is performed to eliminate information redundancy caused by fusion. Thirdly, the modified CMT (MCMT) is constructed to enhance the feature of the primary modality by that of each optimized auxiliary modality feature. Fourthly, Self -Attention Transformer (SAT) is performed on the concatenation result of all the enhanced primary modality features to take full advantage of the common as well complementary properties among them in representing the emotional states. Finally, the enhanced primary modality feature and the optimized auxiliary features are fused by concatenation for emotion recognition. Extensive experimental results on DEAP and WESAD datasets demonstrate that (i) The incongruity does exist among different physiological signals, and the CMT-based auxiliary modality feature optimization strategy can eliminate the incongruity prominently; (ii) The emotion prediction accuracy of the primary modality can be enhanced by the auxiliary modality; (iii) All the key modules in the proposed model, CMT, LRF, and MCMT, contribute to the performance enhancement of the proposed model; iv) The proposed model outperforms State -Of -The -Art (SOTA) models in emotion recognition task.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Incongruity-Aware Cross-Modal Attention for Audio-Visual Fusion in Dimensional Emotion Recognition
    Praveen, R. Gnana
    Alam, Jahangir
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (03) : 444 - 458
  • [2] Context-aware Multimodal Fusion for Emotion Recognition
    Li, Jinchao
    Wang, Shuai
    Chao, Yang
    Liu, Xunying
    Meng, Helen
    [J]. INTERSPEECH 2022, 2022, : 2013 - 2017
  • [3] Multimodal emotion recognition for the fusion of speech and EEG signals
    Ma J.
    Sun Y.
    Zhang X.
    [J]. Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2019, 46 (01): : 143 - 150
  • [4] Multimodal Physiological Signals Fusion for Online Emotion Recognition
    Pan, Tongjie
    Ye, Yalan
    Cai, Hecheng
    Huang, Shudong
    Yang, Yang
    Wang, Guoqing
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5879 - 5888
  • [5] Emotion Recognition from Multimodal Physiological Signals for Emotion Aware Healthcare Systems
    Değer Ayata
    Yusuf Yaslan
    Mustafa E. Kamasak
    [J]. Journal of Medical and Biological Engineering, 2020, 40 : 149 - 157
  • [6] Emotion Recognition from Multimodal Physiological Signals for Emotion Aware Healthcare Systems
    Ayata, Deger
    Yaslan, Yusuf
    Kamasak, Mustafa E.
    [J]. JOURNAL OF MEDICAL AND BIOLOGICAL ENGINEERING, 2020, 40 (02) : 149 - 157
  • [7] Feature-Level Fusion of Multimodal Physiological Signals for Emotion Recognition
    Chen, Jing
    Ru, Bin
    Xu, Lixin
    Moore, Philip
    Su, Yun
    [J]. PROCEEDINGS 2015 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE, 2015, : 395 - 399
  • [8] Multimodal emotion recognition based on the fusion of vision, EEG, ECG, and EMG signals
    Bhatlawande, Shripad
    Pramanik, Sourjadip
    Shilaskar, Swati
    Sole, Swarali
    [J]. INTERNATIONAL JOURNAL OF ELECTRICAL AND COMPUTER ENGINEERING SYSTEMS, 2024, 15 (01) : 41 - 58
  • [9] Emotion-Aware Multimodal Fusion for Meme Emotion Detection
    Sharma, Shivam
    Ramaneswaran, S.
    Akhtar, Md. Shad
    Chakraborty, Tanmoy
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 1800 - 1811
  • [10] COLD Fusion: Calibrated and Ordinal Latent Distribution Fusion for Uncertainty-Aware Multimodal Emotion Recognition
    Tellamekala, Mani Kumar
    Amiriparian, Shahin
    Schuller, Bjorn W.
    Andre, Elisabeth
    Giesbrecht, Timo
    Valstar, Michel
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (02) : 805 - 822