Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis

被引:3
|
作者
Zhang, Rui [1 ,2 ]
Xue, Chengrong [1 ,2 ]
Qi, Qingfu [3 ]
Lin, Liyuan [2 ]
Zhang, Jing [1 ,2 ]
Zhang, Lun [1 ,2 ]
机构
[1] Tianjin Sino German Univ Appl Sci, Sch Software & Commun, Tianjin 300222, Peoples R China
[2] Tianjin Univ Sci & Technol, Coll Elect Informat & Automation, Tianjin 300222, Peoples R China
[3] Gaussian Robot Pte Ltd, Tianjin 200100, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 03期
关键词
multimodal sentiment analysis; bimodal fusion; multi-head attention; EMOTION RECOGNITION; FEATURES;
D O I
10.3390/app13031915
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
The enrichment of social media expression makes multimodal sentiment analysis a research hotspot. However, modality heterogeneity brings great difficulties to effective cross-modal fusion, especially the modality alignment problem and the uncontrolled vector offset during fusion. In this paper, we propose a bimodal multi-head attention network (BMAN) based on text and audio, which adaptively captures the intramodal utterance features and complex intermodal alignment relationships. Specifically, we first set two independent unimodal encoders to extract the semantic features within each modality. Considering that different modalities deserve different weights, we further built a joint decoder to fuse the audio information into the text representation, based on learnable weights to avoid an unreasonable vector offset. The obtained cross-modal representation is used to improve the sentiment prediction performance. Experiments on both the aligned and unaligned CMU-MOSEI datasets show that our model achieves better performance than multiple baselines, and it has outstanding advantages in solving the problem of cross-modal alignment.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Feature Fusion and Multi-head Attention Based Hindi Captioner
    Meghwal, Virendra Kumar
    Mittal, Namita
    Singh, Girdhari
    COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT I, 2024, 2009 : 479 - 487
  • [42] Multi-Head Self-Attention Transformation Networks for Aspect-Based Sentiment Analysis
    Lin, Yuming
    Wang, Chaoqiang
    Song, Hao
    Li, You
    IEEE ACCESS, 2021, 9 : 8762 - 8770
  • [43] Multi-Head Attention with Diversity for Learning Grounded Multilingual Multimodal Representations
    Huang, Po-Yao
    Chang, Xiaojun
    Hauptmann, Alexander
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1461 - 1467
  • [44] Predicting disease genes based on multi-head attention fusion
    Linlin Zhang
    Dianrong Lu
    Xuehua Bi
    Kai Zhao
    Guanglei Yu
    Na Quan
    BMC Bioinformatics, 24
  • [45] Predicting disease genes based on multi-head attention fusion
    Zhang, Linlin
    Lu, Dianrong
    Bi, Xuehua
    Zhao, Kai
    Yu, Guanglei
    Quan, Na
    BMC BIOINFORMATICS, 2023, 24 (01)
  • [46] MULTI-CHANNEL ATTENTIVE GRAPH CONVOLUTIONAL NETWORK WITH SENTIMENT FUSION FOR MULTIMODAL SENTIMENT ANALYSIS
    Xiao, Luwei
    Wu, Xingjiao
    Wu, Wen
    Yang, Jing
    He, Liang
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4578 - 4582
  • [47] Convolutional multi-head self-attention on memory for aspect sentiment classification
    Zhang, Yaojie
    Xu, Bing
    Zhao, Tiejun
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2020, 7 (04) : 1038 - 1044
  • [48] RETRACTED: Sentiment analysis of student feedback using multi-head attention fusion model of word and context embedding for LSTM (Retracted Article)
    Sangeetha, K.
    Prabha, D.
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (03) : 4117 - 4126
  • [49] On the diversity of multi-head attention
    Li, Jian
    Wang, Xing
    Tu, Zhaopeng
    Lyu, Michael R.
    NEUROCOMPUTING, 2021, 454 : 14 - 24
  • [50] EMAT: Efficient feature fusion network for visual tracking via optimized multi-head attention
    Wang, Jun
    Lai, Changwang
    Wang, Yuanyun
    Zhang, Wenshuang
    NEURAL NETWORKS, 2024, 172