Comprehensive Multisource Learning Network for Cross-Subject Multimodal Emotion Recognition

被引:0
|
作者
Chen, Chuangquan [1 ]
Li, Zhencheng [1 ]
Kou, Kit Ian [2 ]
Du, Jie [3 ]
Li, Chen [4 ]
Wang, Hongtao [1 ]
Vong, Chi-Man [5 ]
机构
[1] Wuyi Univ, Sch Elect & Informat Engn, Jiangmen 529020, Peoples R China
[2] Univ Macau, Dept Math, Taipa 999078, Macao, Peoples R China
[3] Shenzhen Univ, Med Sch, Sch Biomed Engn, Shenzhen 518060, Peoples R China
[4] Foshan Univ, Sch Math & Big Data, Foshan 528225, Peoples R China
[5] Univ Macau, Dept Comp & Informat Sci, Taipa 999078, Macao, Peoples R China
基金
中国国家自然科学基金;
关键词
Multisource learning; multimodal emotion recognition; cross-subject; low-rank multimodal fusion; domain generalization;
D O I
10.1109/TETCI.2024.3406422
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Electroencephalography (EEG) signals and eye movement signals, which represent internal physiological responses and external subconscious behaviors, respectively, have been shown to be reliable indicators for recognizing emotions. However, integrating these two modalities across multiple subjects presents several challenges: 1) designing a robust consistency metric that balances the consistency and divergences between heterogeneous modalities across multiple subjects; 2) simultaneously considering intra-modality and inter-modality information across multiple subjects; and 3) overcoming individual differences among multiple subjects and generating subject-invariant representations of the multimodal fused features. To address these challenges associated with multisource data (i.e., multiple modalities and subjects), we propose a novel comprehensive multisource learning network (CMSLNet) for cross-subject multimodal emotion recognition. Specifically, an instance-level adaptive robust consistency metric is first designed to better align the information between EEG signals and eye movement signals, identifying their consistency and divergences across various emotions. Subsequently, an attentive low-rank multimodal fusion (Att-LMF) method is developed to account for individual differences and dynamically learn intra-modality and inter-modality information, resulting in highly discriminative fused features. Finally, domain generalization is utilized to extract subject-invariant representations of the fused features, thus adapting to new subjects and enhancing the model's generalization. Through these elaborate designs, CMSLNet effectively incorporates the information from multisource data, thus significantly improving the accuracy and reliability of emotion recognition. Extensive experiments on two public datasets demonstrate the superior performance of CMSLNet. CMSLNet achieves high accuracies of 83.15% on the SEED-IV dataset and 87.32% on the SEED-V dataset, surpassing the state-of-the-art methods by 3.62% and 4.60%, respectively.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Multisource Transfer Learning for Cross-Subject EEG Emotion Recognition
    Li, Jinpeng
    Qiu, Shuang
    Shen, Yuan-Yuan
    Liu, Cheng-Lin
    He, Huiguang
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) : 3281 - 3293
  • [2] Cross-Subject Multimodal Emotion Recognition Based on Hybrid Fusion
    Cimtay, Yucel
    Ekmekcioglu, Erhan
    Caglar-Ozhan, Seyma
    [J]. IEEE ACCESS, 2020, 8 : 168865 - 168878
  • [3] Multisource Associate Domain Adaptation for Cross-Subject and Cross-Session EEG Emotion Recognition
    She, Qingshan
    Zhang, Chenqi
    Fang, Feng
    Ma, Yuliang
    Zhang, Yingchun
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [4] MASS: A Multisource Domain Adaptation Network for Cross-Subject Touch Gesture Recognition
    Li, Yun-Kai
    Meng, Qing-Hao
    Wang, Ya-Xin
    Yang, Tian-Hao
    Hou, Hui-Rang
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (03) : 3099 - 3108
  • [5] Hybrid transfer learning strategy for cross-subject EEG emotion recognition
    Lu, Wei
    Liu, Haiyan
    Ma, Hua
    Tan, Tien-Ping
    Xia, Lingnan
    [J]. FRONTIERS IN HUMAN NEUROSCIENCE, 2023, 17
  • [6] Contrastive Learning of Subject-Invariant EEG Representations for Cross-Subject Emotion Recognition
    Shen, Xinke
    Liu, Xianggen
    Hu, Xin
    Zhang, Dan
    Song, Sen
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2496 - 2511
  • [7] Domain Adaptation for Cross-Subject Emotion Recognition by Subject Clustering
    Liu, Jin
    Shen, Xinke
    Song, Sen
    Zhang, Dan
    [J]. 2021 10TH INTERNATIONAL IEEE/EMBS CONFERENCE ON NEURAL ENGINEERING (NER), 2021, : 904 - 908
  • [8] Cross-Subject EEG-Based Emotion Recognition via Semisupervised Multisource Joint Distribution Adaptation
    Jimenez-Guarneros, Magdiel
    Fuentes-Pineda, Gibran
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [9] A Novel Experiment Setting for Cross-subject Emotion Recognition
    Hu, Hao-Yi
    Zhao, Li-Ming
    Liu, Yu-Zhong
    Li, Hua-Liang
    Lu, Bao-Liang
    [J]. 2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 6416 - 6419
  • [10] Evolutionary Ensemble Learning for EEG-Based Cross-Subject Emotion Recognition
    Zhang, Hanzhong
    Zuo, Tienyu
    Chen, Zhiyang
    Wang, Xin
    Sun, Poly Z. H.
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (07) : 3872 - 3881