Causal Inference for Modality Debiasing in Multimodal Emotion Recognition

被引:0
|
作者
Kim, Juyeon [1 ]
Hong, Juyoung [1 ]
Choi, Yukyung [1 ]
机构
[1] Sejong Univ, Dept Convergence Engn Intelligent Drone, Seoul 05006, South Korea
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 23期
关键词
emotion recognition; multimodal learning; causal inference;
D O I
10.3390/app142311397
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Multimodal emotion recognition (MER) aims to enhance the understanding of human emotions by integrating visual, auditory, and textual modalities. However, previous MER approaches often depend on a dominant modality rather than considering all modalities, leading to poor generalization. To address this, we propose Causal Inference in Multimodal Emotion Recognition (CausalMER), which leverages counterfactual reasoning and causal graphs to capture relationships between modalities and reduce direct modality effects contributing to bias. This allows CausalMER to make unbiased predictions while being easily applied to existing MER methods in a model-agnostic manner, without requiring any architectural modifications. We evaluate CausalMER on the IEMOCAP and CMU-MOSEI datasets, widely used benchmarks in MER, and compare it with existing methods. On the IEMOCAP dataset with the MulT backbone, CausalMER achieves an average accuracy of 83.4%. On the CMU-MOSEI dataset, the average accuracies with MulT, PMR, and DMD backbones are 50.1%, 48.8%, and 48.8%, respectively. Experimental results demonstrate that CausalMER is robust in missing modality scenarios, as shown by its low standard deviation in performance drop gaps. Additionally, we evaluate modality contributions and show that CausalMER achieves balanced contributions from each modality, effectively mitigating direct biases from individual modalities.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Simplifying Multimodal Emotion Recognition with Single Eye Movement Modality
    Yan, Xu
    Zhao, Li-Ming
    Lu, Bao-Liang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1057 - 1063
  • [2] A Survey on Debiasing Recommendation Based on Causal Inference
    Yang, Xin-Xin
    Liu, Zhen
    Lu, Si-Bo
    Yuan, Ya-Fan
    Sun, Yong-Qi
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (10): : 2307 - 2332
  • [3] Progressive Modality Reinforcement for Human Multimodal Emotion Recognition from Unaligned Multimodal Sequences
    Lv, Fengmao
    Chen, Xiang
    Huang, Yanyong
    Duan, Lixin
    Lin, Guosheng
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2554 - 2562
  • [4] Modality-Transferable Emotion Embeddings for Low-Resource Multimodal Emotion Recognition
    Dai, Wenliang
    Liu, Zihan
    Yu, Tiezheng
    Fung, Pascale
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 269 - 280
  • [5] Temporal Relation Inference Network for Multimodal Speech Emotion Recognition
    Dong, Guan-Nan
    Pun, Chi-Man
    Zhang, Zheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (09) : 6472 - 6485
  • [6] Debiasing Multimodal Models via Causal Information Minimization
    Patil, Vaidehi
    Maharana, Adyasha
    Bansal, Mohit
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4108 - 4123
  • [7] Multimodal Emotion Recognition with Modality-Pairwise Unsupervised Contrastive Loss
    Franceschini, Riccardo
    Fini, Enrico
    Beyan, Cigdem
    Conti, Alessandro
    Arrigoni, Federica
    Ricci, Elisa
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2589 - 2596
  • [8] Revisiting Disentanglement and Fusion on Modality and Context in Conversational Multimodal Emotion Recognition
    Li, Bobo
    Fei, Hao
    Liao, Lizi
    Zhao, Yu
    Teng, Chong
    Chua, Tat-Seng
    Ji, Donghong
    Li, Fei
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5923 - 5934
  • [9] "Select language, modality or put on a mask!" Experiments with multimodal emotion recognition
    Bujnowski, Pawel
    Kuzma, Bartlomiej
    Paziewski, Bartlomiej
    Rutkowski, Jacek
    Marhula, Joanna
    Bordzicka, Zuzanna
    Andruszkiewicz, Piotr
    INTERSPEECH 2023, 2023, : 672 - 673
  • [10] Multimodal Emotion Recognition with Vision-language Prompting and Modality Dropout
    Qi, Anbin
    Liu, Zhongliang
    Zhou, Xinyong
    Xiao, Jinba
    Zhang, Fengrun
    Gan, Qi
    Tao, Ming
    Zhang, Gaozheng
    Zhang, Lu
    PROCEEDINGS OF THE 2ND INTERNATIONAL WORKSHOP ON MULTIMODAL AND RESPONSIBLE AFFECTIVE COMPUTING, MRAC 2024, 2024, : 49 - 53