Multi-level Multi-task representation learning with adaptive fusion for multimodal sentiment analysis

被引:0
|
作者
Chuanbo Zhu [1 ]
Min Chen [2 ]
Haomin Li [3 ]
Sheng Zhang [1 ]
Han Liang [1 ]
Chao Sun [1 ]
Yifan Liu [1 ]
Jincai Chen [1 ]
机构
[1] Huazhong University of Science and Technology,Wuhan National Laboratory for Optoelectronics
[2] South China University of Technology,School of Computer Science and Engineering
[3] Pazhou Laboratory,School of Computer Science and Technology
[4] Huazhong University of Science and Technology,Key Laboratory of Information Storage System
[5] Ministry of Education of China,undefined
关键词
Multimodal sentiment analysis; Multimodal adaptive fusion; Multi-level representation; Multi-task learning;
D O I
10.1007/s00521-024-10678-1
中图分类号
学科分类号
摘要
Multimodal sentiment analysis is an active task in multimodal intelligence, which aims to compute the user’s sentiment tendency from multimedia data. Generally, each modality is a specific and necessary perspective to express human sentiment, providing complementary and consensus information unavailable in a single modality. Nevertheless, the heterogeneous multimedia data often contain inconsistent and conflicting sentiment semantics that limits the model performance. In this work, we propose a Multi-level Multi-task Representation Learning with Adaptive Fusion (MuReLAF) network to bridge the semantic gap among different modalities. Specifically, we design a modality adaptive fusion block to adjust modality contributions dynamically. Besides, we build a multi-level multimodal representations framework to obtain modality-specific and modality-shared semantics by the multi-task learning strategy, where modality-specific semantics contain complementary information and modality-shared semantics include consensus information. Extensive experiments are conducted on four publicly available datasets: MOSI, MOSEI, SIMS, and SIMSV2(s), demonstrating that our model exhibits superior or comparable performance to state-of-the-art models. The achieved accuracies are 86.28%, 86.07%, 84.46%, and 82.78%, respectively, showcasing improvements of 0.82%, 0.84%, 1.75%, and 1.83%. Further analyses also indicate the effectiveness of our model in sentiment analysis.
引用
收藏
页码:1491 / 1508
页数:17
相关论文
共 50 条
  • [31] Multi-Task Network Representation Learning
    Xie, Yu
    Jin, Peixuan
    Gong, Maoguo
    Zhang, Chen
    Yu, Bin
    FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [32] Dual adaptive learning multi-task multi-view for graph network representation learning
    Han, Beibei
    Wei, Yingmei
    Wang, Qingyong
    Wan, Shanshan
    NEURAL NETWORKS, 2023, 162 : 297 - 308
  • [33] Learning Modality-Specific Representations with Self-Supervised Multi-Task Learning for Multimodal Sentiment Analysis
    Yu, Wenmeng
    Xu, Hua
    Yuan, Ziqi
    Wu, Jiele
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10790 - 10797
  • [34] A Multi-step Attention and Multi-level Structure Network for Multimodal Sentiment Analysis
    Zhang, Chuanlei
    Zhao, Hongwei
    Wang, Bo
    Wang, Wei
    Ke, Ting
    Li, Jianrong
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 723 - 735
  • [35] Learning Modality-Complementary and Eliminating-Redundancy Representations with Multi-Task Learning for Multimodal Sentiment Analysis
    Zhao, Xiaowei
    Miao, Xinyu
    Xu, Xiujuan
    Liu, Yu
    Cao, Yifei
    2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,
  • [36] Multimodal sentiment analysis model based on multi-task learning and stacked cross-modal Transformer
    Chen Q.-H.
    Sun J.-J.
    Lou Y.-B.
    Fang Z.-J.
    Zhejiang Daxue Xuebao (Gongxue Ban)/Journal of Zhejiang University (Engineering Science), 2023, 57 (12): : 2421 - 2429
  • [37] ADAPTIVE AND ROBUST MULTI-TASK LEARNING
    Duan, Yaqi
    Wang, Kaizheng
    ANNALS OF STATISTICS, 2023, 51 (05): : 2015 - 2039
  • [38] Attention Based Shared Representation for Multi-task Stance Detection and Sentiment Analysis
    Chauhan, Dushyant Singh
    Kumar, Rohan
    Ekbal, Asif
    NEURAL INFORMATION PROCESSING, ICONIP 2019, PT V, 2019, 1143 : 661 - 669
  • [39] A Multi-Task Learning Approach to Improve Sentiment Analysis with Explicit Recommendation
    Habimana, Olivier
    Li, Yuhua
    Li, Ruixuan
    Gu, Xiwu
    Peng, Yuqi
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [40] Sentiment Analysis and Sarcasm Detection using Deep Multi-Task Learning
    Yik Yang Tan
    Chee-Onn Chow
    Jeevan Kanesan
    Joon Huang Chuah
    YongLiang Lim
    Wireless Personal Communications, 2023, 129 : 2213 - 2237