SKEAFN: Sentiment Knowledge Enhanced Attention Fusion Network for multimodal sentiment analysis

被引:15
|
作者
Zhu, Chuanbo [1 ]
Chen, Min [2 ,3 ]
Zhang, Sheng [1 ]
Sun, Chao [1 ]
Liang, Han [1 ]
Liu, Yifan [1 ]
Chen, Jincai [1 ,2 ,4 ]
机构
[1] Huazhong Univ Sci & Technol, Wuhan Natl Lab Optoelect, Wuhan 430074, Hubei, Peoples R China
[2] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, Wuhan 430074, Hubei, Peoples R China
[3] Huazhong Univ Sci & Technol, Embedded & Pervas Comp EP Lab, Wuhan 430074, Hubei, Peoples R China
[4] Minist Educ China, Engn Res Ctr Data Storage Syst & Technol, Key Lab Informat Storage Syst, Wuhan 430074, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-view learning; Multiple feature fusion; Multimodal sentiment analysis; External knowledge; Multi-head attention;
D O I
10.1016/j.inffus.2023.101958
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal sentiment analysis is an active research field that aims to recognize the user's sentiment information from multimodal data. The primary challenge in this field is to develop a high-quality fusion framework that effectively addresses the heterogeneity among different modalities. However, prior research has primarily concentrated on intermodal interactions while neglecting the semantic sentiment information conveyed by words in the text modality. In this paper, we propose the Sentiment Knowledge Enhanced Attention Fusion Network (SKEAFN), a novel end-to-end fusion network that enhances multimodal fusion by incorporating additional sentiment knowledge representations from an external knowledge base. Firstly, we construct an external knowledge enhancement module to acquire additional representations for the text modality. Then, we design a text-guided interaction module that facilitates the interaction between text and the visual/acoustic modality. Finally, we propose a feature-wised attention fusion module that achieves multimodal fusion by dynamically adjusting the weights of the additional and each modality's representations. We evaluate our method on three challenging multimodal sentiment analysis datasets: CMU-MOSI, CMU-MOSEI, and Twitter2019. The experiment results demonstrate that our model significantly outperforms the state-of-the-art models. The source code is publicly available at https://github.com/doubibobo/SKEAFN.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Attention fusion network for multimodal sentiment analysis
    Yuanyi Luo
    Rui Wu
    Jiafeng Liu
    Xianglong Tang
    Multimedia Tools and Applications, 2024, 83 : 8207 - 8217
  • [2] Attention fusion network for multimodal sentiment analysis
    Luo, Yuanyi
    Wu, Rui
    Liu, Jiafeng
    Tang, Xianglong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (03) : 8207 - 8217
  • [3] Graph Reconstruction Attention Fusion Network for Multimodal Sentiment Analysis
    Hu, Ronglong
    Yi, Jizheng
    Chen, Lijiang
    Jin, Ze
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2025, 21 (01) : 297 - 306
  • [4] Gated attention fusion network for multimodal sentiment classification
    Du, Yongping
    Liu, Yang
    Peng, Zhi
    Jin, Xingnan
    KNOWLEDGE-BASED SYSTEMS, 2022, 240
  • [5] Multimodal Sentiment Analysis Based on Attention Mechanism and Tensor Fusion Network
    Zhang, Kang
    Geng, Yushui
    Zhao, Jing
    Li, Wenxiao
    Liu, Jianxin
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 1473 - 1477
  • [6] TETFN: A text enhanced transformer fusion network for multimodal sentiment analysis
    Wang, Di
    Guo, Xutong
    Tian, Yumin
    Liu, Jinhui
    He, LiHuo
    Luo, Xuemei
    PATTERN RECOGNITION, 2023, 136
  • [7] A multimodal fusion network with attention mechanisms for visual-textual sentiment analysis
    Gan, Chenquan
    Fu, Xiang
    Feng, Qingdong
    Zhu, Qingyi
    Cao, Yang
    Zhu, Ye
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 242
  • [8] Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis
    Zhang, Rui
    Xue, Chengrong
    Qi, Qingfu
    Lin, Liyuan
    Zhang, Jing
    Zhang, Lun
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [9] Sentiment analysis of social media comments based on multimodal attention fusion network
    Liu, Ziyu
    Yang, Tao
    Chen, Wen
    Chen, Jiangchuan
    Li, Qinru
    Zhang, Jun
    APPLIED SOFT COMPUTING, 2024, 164
  • [10] Sentiment Knowledge Enhanced Self-supervised Learning for Multimodal Sentiment Analysis
    Qian, Fan
    Han, Jiqing
    He, Yongjun
    Zheng, Tieran
    Zheng, Guibin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 12966 - 12978