Multi-level Multi-task representation learning with adaptive fusion for multimodal sentiment analysis

被引:0
|
作者
Chuanbo Zhu [1 ]
Min Chen [2 ]
Haomin Li [3 ]
Sheng Zhang [1 ]
Han Liang [1 ]
Chao Sun [1 ]
Yifan Liu [1 ]
Jincai Chen [1 ]
机构
[1] Huazhong University of Science and Technology,Wuhan National Laboratory for Optoelectronics
[2] South China University of Technology,School of Computer Science and Engineering
[3] Pazhou Laboratory,School of Computer Science and Technology
[4] Huazhong University of Science and Technology,Key Laboratory of Information Storage System
[5] Ministry of Education of China,undefined
关键词
Multimodal sentiment analysis; Multimodal adaptive fusion; Multi-level representation; Multi-task learning;
D O I
10.1007/s00521-024-10678-1
中图分类号
学科分类号
摘要
Multimodal sentiment analysis is an active task in multimodal intelligence, which aims to compute the user’s sentiment tendency from multimedia data. Generally, each modality is a specific and necessary perspective to express human sentiment, providing complementary and consensus information unavailable in a single modality. Nevertheless, the heterogeneous multimedia data often contain inconsistent and conflicting sentiment semantics that limits the model performance. In this work, we propose a Multi-level Multi-task Representation Learning with Adaptive Fusion (MuReLAF) network to bridge the semantic gap among different modalities. Specifically, we design a modality adaptive fusion block to adjust modality contributions dynamically. Besides, we build a multi-level multimodal representations framework to obtain modality-specific and modality-shared semantics by the multi-task learning strategy, where modality-specific semantics contain complementary information and modality-shared semantics include consensus information. Extensive experiments are conducted on four publicly available datasets: MOSI, MOSEI, SIMS, and SIMSV2(s), demonstrating that our model exhibits superior or comparable performance to state-of-the-art models. The achieved accuracies are 86.28%, 86.07%, 84.46%, and 82.78%, respectively, showcasing improvements of 0.82%, 0.84%, 1.75%, and 1.83%. Further analyses also indicate the effectiveness of our model in sentiment analysis.
引用
收藏
页码:1491 / 1508
页数:17
相关论文
共 50 条
  • [41] Sentiment Analysis and Sarcasm Detection using Deep Multi-Task Learning
    Tan, Yik Yang
    Chow, Chee-Onn
    Kanesan, Jeevan
    Chuah, Joon Huang
    Lim, YongLiang
    WIRELESS PERSONAL COMMUNICATIONS, 2023, 129 (03) : 2213 - 2237
  • [42] Text-guided multi-level interaction and multi-scale spatial-memory fusion for multimodal sentiment analysis
    He, Xiaojiang
    Fang, Yanjie
    Xiang, Nan
    Li, Zuhe
    Wang, Qiufeng
    Yang, Chenguang
    Wang, Hao
    Pan, Yushan
    NEUROCOMPUTING, 2025, 626
  • [43] Multi-view representation learning in multi-task scene
    Run-kun Lu
    Jian-wei Liu
    Si-ming Lian
    Xin Zuo
    Neural Computing and Applications, 2020, 32 : 10403 - 10422
  • [44] Multi-view representation learning in multi-task scene
    Lu, Run-kun
    Liu, Jian-wei
    Lian, Si-ming
    Zuo, Xin
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (14): : 10403 - 10422
  • [45] Single-image super-resolution via dynamic multi-task learning and multi-level mutual feature fusion
    Zhang, Yaofang
    Fang, Yuchun
    Ran, Qicai
    Wu, Jiahua
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (03)
  • [46] Optimistic Rates for Multi-Task Representation Learning
    Watkins, Austin
    Ullah, Enayat
    Thanh Nguyen-Tang
    Arora, Raman
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [47] Multi-Task Learning for Sentiment Analysis with Hard-Sharing and Task Recognition Mechanisms
    Zhang, Jian
    Yan, Ke
    Mo, Yuchang
    INFORMATION, 2021, 12 (05)
  • [48] Multi-Task Network Combing Multi-Level Information for Object Localization
    Tian, Yan
    Wang, Huiyan
    Wang, Xun
    Huang, Gang
    Zhang, Guofeng
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2017, 29 (07): : 1275 - 1282
  • [49] Task Adaptive Parameter Sharing for Multi-Task Learning
    Wallingford, Matthew
    Li, Hao
    Achille, Alessandro
    Ravichandran, Avinash
    Fowlkes, Charless
    Bhotika, Rahul
    Soatto, Stefano
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7551 - 7560
  • [50] PURE: Personality-Coupled Multi-Task Learning Framework for Aspect-Based Multimodal Sentiment Analysis
    Zhang, Puning
    Fu, Miao
    Zhao, Rongjian
    Zhang, Hongbin
    Luo, Changchun
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (01) : 462 - 477