A Two-Stage Multi-Modal Multi-Label Emotion Recognition Decision System Based on GCN

被引:0
|
作者
Wu, Weiwei [1 ]
Chen, Daomin [2 ]
Li, Qingping [1 ]
机构
[1] Zhejiang Yuying Coll Vocat Technol, Hangzhou, Peoples R China
[2] Guangdong Univ Sci & Technol, Guangzhou, Peoples R China
关键词
Multiple Modalities; Multi-Label Classification; Graph Convolutional Networks; Emotion Detection;
D O I
10.4018/IJDSST.352398
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Compared with single-modal methods, emotion recognition research is increasingly focusing on the use of multi-modal methods to improve accuracy. Despite the advantages of multimodality, challenges such as feature fusion and redundancy remain. In this study, we propose a multi-modal multi-label emotion recognition decision system based on graph convolution. Our approach utilizes text, speech, and video data for feature extraction, while combining tag attention to capture fine-grained modal dependencies. The two-stage feature reconstruction module facilitates complementary feature fusion while preserving mode-specific information. Emotional decisions are made using a fully connected layer to optimize performance without adding complexity to the model. Experimental results on IEMOCAP, CMU-MOSEI and MELD datasets show that our algorithm has higher accuracy than existing models, highlighting the effectiveness and innovation of our proposed algorithm.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Multi-Modal Fusion Emotion Recognition Based on HMM and ANN
    Xu, Chao
    Cao, Tianyi
    Feng, Zhiyong
    Dong, Caichao
    CONTEMPORARY RESEARCH ON E-BUSINESS TECHNOLOGY AND STRATEGY, 2012, 332 : 541 - 550
  • [32] A Deep Multi-Modal CNN for Multi-Instance Multi-Label Image Classification
    Song, Lingyun
    Liu, Jun
    Qian, Buyue
    Sun, Mingxuan
    Yang, Kuan
    Sun, Meng
    Abbas, Samar
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (12) : 6025 - 6038
  • [33] Multi-Modal Multi-Instance Multi-Label Learning with Graph Convolutional Network
    Hang, Cheng
    Wang, Wei
    Zhan, De-Chuan
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [34] A two-stage multi-view partial multi-label learning for enhanced disambiguation
    Wang, Ziyi
    Xu, Yitian
    KNOWLEDGE-BASED SYSTEMS, 2024, 293
  • [35] Partial Modal Conditioned GANs for Multi-modal Multi-label Learning with Arbitrary Modal-Missing
    Zhang, Yi
    Shen, Jundong
    Zhang, Zhecheng
    Wang, Chongjun
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT II, 2021, 12682 : 413 - 428
  • [36] Micro-video multi-label classification method based on multi-modal feature encoding
    Jing P.
    Li Y.
    Su Y.
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2022, 49 (04): : 109 - 117
  • [37] TRANSFORMER-BASED MULTI-MODAL LEARNING FOR MULTI-LABEL REMOTE SENSING IMAGE CLASSIFICATION
    Hoffmann, David Sebastian
    Clasen, Kai Norman
    Demir, Begum
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 4891 - 4894
  • [38] MIC: Breast Cancer Multi-label Diagnostic Framework Based on Multi-modal Fusion Interaction
    Chen, Ziyan
    Yi, Sanli
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025,
  • [39] DriveSense: A Multi-modal Emotion Recognition and Regulation System for a Car Driver
    Zhu, Lei
    Zhong, Zhinan
    Dai, Wan
    Chen, Yunfei
    Zhang, Yan
    Chen, Mo
    HCI IN MOBILITY, TRANSPORT, AND AUTOMOTIVE SYSTEMS, MOBITAS 2024, PT I, 2024, 14732 : 82 - 97
  • [40] Multi-modal Contextual Prompt Learning for Multi-label Classification with Partial Labels
    Wang, Rui
    Pan, Zhengxin
    Wu, Fangyu
    Lv, Yifan
    Zhang, Bailing
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 517 - 524