Scanning, attention, and reasoning multimodal content for sentiment analysis

被引:5
|
作者
Liu, Yun [1 ]
Li, Zhoujun [2 ]
Zhou, Ke [1 ]
Zhang, Leilei [1 ]
Li, Lang [1 ]
Tian, Peng [1 ]
Shen, Shixun [1 ]
机构
[1] Moutai Inst, Dept Automat, Renhuai 564507, Guizhou Provinc, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodal sentiment analysis; Attention; Reasoning; FUSION;
D O I
10.1016/j.knosys.2023.110467
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rise of social networks has provided people with platforms to display their lives and emotions, often in multimodal forms such as images and descriptive texts. Capturing the emotions embedded in the multimodal content of social networks involves great research challenges and practical values. Existing methods usually make sentiment predictions based on a single-round reasoning process with multimodal attention networks, however, this may be insufficient for tasks that require deep understanding and complex reasoning. To effectively comprehend multimodal content and predict the correct sentiment tendencies, we propose the Scanning, Attention, and Reasoning (SAR) model for multimodal sentiment analysis. Specifically, a perceptual scanning model is designed to roughly perceive the image and text content, as well as the intrinsic correlation between them. To deeply understand the complementary features between images and texts, an intensive attention model is proposed for cross-modal feature association learning. The multimodal joint features from the scanning and attention models are fused together as the representation of a multimodal node in the social network. A heterogeneous reasoning model implemented with a graph neural network is constructed to capture the influence of network communication in social networks and make sentiment predictions. Extensive experiments conducted on three benchmark datasets confirm the effectiveness and superiority of our model compared with state-of-the-art methods.(c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Multimodal Sentiment Analysis Model Integrating Multi-features and Attention Mechanism
    Lyu X.
    Tian C.
    Zhang L.
    Du Y.
    Zhang X.
    Cai Z.
    Data Analysis and Knowledge Discovery, 2024, 8 (05) : 91 - 101
  • [32] Multimodal GRU with directed pairwise cross-modal attention for sentiment analysis
    Zhenkai Qin
    Qining Luo
    Zhidong Zang
    Hongpeng Fu
    Scientific Reports, 15 (1)
  • [33] Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis
    Zhang, Rui
    Xue, Chengrong
    Qi, Qingfu
    Lin, Liyuan
    Zhang, Jing
    Zhang, Lun
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [34] Multimodal Sentiment Analysis Based on a Cross-Modal Multihead Attention Mechanism
    Deng, Lujuan
    Liu, Boyi
    Li, Zuhe
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 78 (01): : 1157 - 1170
  • [35] A Multimodal Sentiment Analysis Approach Based on a Joint Chained Interactive Attention Mechanism
    Qiu, Keyuan
    Zhang, Yingjie
    Zhao, Jiaxu
    Zhang, Shun
    Wang, Qian
    Chen, Feng
    ELECTRONICS, 2024, 13 (10)
  • [36] EGSRNet: Emotion-Label Guiding and Similarity Reasoning Network for Multimodal Sentiment Analysis
    Zhan, Chunlan
    Qian, Wenhua
    Liu, Peng
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 364 - 378
  • [37] Context-Dependent Multimodal Sentiment Analysis Based on a Complex Attention Mechanism
    Deng, Lujuan
    Liu, Boyi
    Li, Zuhe
    Ma, Jiangtao
    Li, Hanbing
    ELECTRONICS, 2023, 12 (16)
  • [38] Sentiment analysis of social media comments based on multimodal attention fusion network
    Liu, Ziyu
    Yang, Tao
    Chen, Wen
    Chen, Jiangchuan
    Li, Qinru
    Zhang, Jun
    APPLIED SOFT COMPUTING, 2024, 164
  • [39] Deep Modular Co-Attention Shifting Network for Multimodal Sentiment Analysis
    Shi, Piao
    Hu, Min
    Shi, Xuefeng
    Ren, Fuji
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (04)
  • [40] A cognitive brain model for multimodal sentiment analysis based on attention neural networks
    Li, Yuanqing
    Zhang, Ke
    Wang, Jingyu
    Gao, Xinbo
    NEUROCOMPUTING, 2021, 430 : 159 - 173