Causal Inference for Modality Debiasing in Multimodal Emotion Recognition

被引:0
|
作者
Kim, Juyeon [1 ]
Hong, Juyoung [1 ]
Choi, Yukyung [1 ]
机构
[1] Sejong Univ, Dept Convergence Engn Intelligent Drone, Seoul 05006, South Korea
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 23期
关键词
emotion recognition; multimodal learning; causal inference;
D O I
10.3390/app142311397
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Multimodal emotion recognition (MER) aims to enhance the understanding of human emotions by integrating visual, auditory, and textual modalities. However, previous MER approaches often depend on a dominant modality rather than considering all modalities, leading to poor generalization. To address this, we propose Causal Inference in Multimodal Emotion Recognition (CausalMER), which leverages counterfactual reasoning and causal graphs to capture relationships between modalities and reduce direct modality effects contributing to bias. This allows CausalMER to make unbiased predictions while being easily applied to existing MER methods in a model-agnostic manner, without requiring any architectural modifications. We evaluate CausalMER on the IEMOCAP and CMU-MOSEI datasets, widely used benchmarks in MER, and compare it with existing methods. On the IEMOCAP dataset with the MulT backbone, CausalMER achieves an average accuracy of 83.4%. On the CMU-MOSEI dataset, the average accuracies with MulT, PMR, and DMD backbones are 50.1%, 48.8%, and 48.8%, respectively. Experimental results demonstrate that CausalMER is robust in missing modality scenarios, as shown by its low standard deviation in performance drop gaps. Additionally, we evaluate modality contributions and show that CausalMER achieves balanced contributions from each modality, effectively mitigating direct biases from individual modalities.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Emotion Recognition using Multimodal Features
    Zhao, Jinming
    Chen, Shizhe
    Wang, Shuai
    Jin, Qin
    2018 FIRST ASIAN CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII ASIA), 2018,
  • [42] A Multimodal Dataset for Mixed Emotion Recognition
    Yang, Pei
    Liu, Niqi
    Liu, Xinge
    Shu, Yezhi
    Ji, Wenqi
    Ren, Ziqi
    Sheng, Jenny
    Yu, Minjing
    Yi, Ran
    Zhang, Dan
    Liu, Yong-Jin
    SCIENTIFIC DATA, 2024, 11 (01)
  • [43] Multimodal approaches for emotion recognition: A survey
    Sebe, N
    Cohen, I
    Gevers, T
    Huang, TS
    INTERNET IMAGING VI, 2005, 5670 : 56 - 67
  • [44] Age Differences in Emotion Recognition: A Question of Modality?
    Wieck, Cornelia
    Kunzmann, Ute
    PSYCHOLOGY AND AGING, 2017, 32 (05) : 401 - 411
  • [45] Modality encoded latent dataset for emotion recognition
    Mert, Ahmet
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 79
  • [46] Multimodal Emotion Recognition Based on the Decoupling of Emotion and Speaker Information
    Gajsek, Rok
    Struc, Vitomir
    Mihelic, France
    TEXT, SPEECH AND DIALOGUE, 2010, 6231 : 275 - 282
  • [47] Contrastive Learning Based Modality-Invariant Feature Acquisition for Robust Multimodal Emotion Recognition With Missing Modalities
    Liu, Rui
    Zuo, Haolin
    Lian, Zheng
    Schuller, Bjorn W.
    Li, Haizhou
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (04) : 1856 - 1873
  • [48] CMOT: Cross-Modality Optimal Transport for multimodal inference
    Sayali Anil Alatkar
    Daifeng Wang
    Genome Biology, 24
  • [49] CMOT: Cross-Modality Optimal Transport for multimodal inference
    Alatkar, Sayali Anil
    Wang, Daifeng
    GENOME BIOLOGY, 2023, 24 (01)
  • [50] Comparing Recognition Performance and Robustness of Multimodal Deep Learning Models for Multimodal Emotion Recognition
    Liu, Wei
    Qiu, Jie-Lin
    Zheng, Wei-Long
    Lu, Bao-Liang
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (02) : 715 - 729