Image-Text Multimodal Sentiment Analysis Framework of Assamese News Articles Using Late Fusion

被引:9
|
作者
Das, Ringki [1 ]
Singh, Thoudam Doren [1 ]
机构
[1] Natl Inst Technol Silchar, Dept Comp Sci & Engn, Silchar 788010, Assam, India
关键词
Multimodal sentiment analysis; low resource language; caption generation; machine learning classifier; late fusion;
D O I
10.1145/3584861
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Before the arrival of the web as a corpus, people detected positive and negative news based on the understanding of the textual content from physical newspaper rather than an automatic identification approach from readily available e-newspapers. Thus, the earlier sentiment analysis approach is based on unimodal data, and less effort is paid to the multimodal data. However, the presence of multimodal information helps us to get a clearer understanding of the sentiment. To the best of our knowledge, less work has been introduced on the image-text multimodal sentiment analysis framework of Assamese, a low-resource Indian language mostly spoken in the northeast part of India. We built an Assamese news articles dataset consisting of news text and associated images and one image caption to conduct an experimental study. Focusing on important words and discriminative regions of the images mostly related to sentiment, two individual unimodal such as textual and visual models are proposed. The visual model is developed using an encoder-decoder-based image caption generation system. An image-text multimodal approach is proposed to explore the internal correlation between textual and visual features for joint sentiment classification. Finally, we propose the multimodal sentiment analysis framework, i.e., Textual Visual Multimodal Fusion, by employing a late fusion scheme to merge the three different modalities for the final sentiment prediction. Experimental results conducted on the Assamese dataset built in-house demonstrate that the contextual integration of multimodal features delivers better performance than unimodal features.
引用
收藏
页数:30
相关论文
共 50 条
  • [31] Sentiment analysis of financial news articles using performance indicators
    Srikumar Krishnamoorthy
    [J]. Knowledge and Information Systems, 2018, 56 : 373 - 394
  • [32] Sentiment analysis of financial news articles using performance indicators
    Krishnamoorthy, Srikumar
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2018, 56 (02) : 373 - 394
  • [33] A Sentimental Prompt Framework with Visual Text Encoder for Multimodal Sentiment Analysis
    Huang, Shizhou
    Xu, Bo
    Li, Changqun
    Ye, Jiabo
    Lin, Xin
    [J]. PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 638 - 646
  • [34] An Image-Text Sentiment Analysis Method Using Multi-Channel Multi-Modal Joint Learning
    Gong, Lianting
    He, Xingzhou
    Yang, Jianzhong
    [J]. APPLIED ARTIFICIAL INTELLIGENCE, 2024, 38 (01)
  • [35] Multimodal Sentiment Analysis using Audio and Text for Crime Detection
    Boukabous, Mohammed
    Azizi, Mostafa
    [J]. 2022 2ND INTERNATIONAL CONFERENCE ON INNOVATIVE RESEARCH IN APPLIED SCIENCE, ENGINEERING AND TECHNOLOGY (IRASET'2022), 2022, : 803 - 807
  • [36] Image-Text Sentiment Analysis Via Context Guided Adaptive Fine-Tuning Transformer
    Xiao, Xingwang
    Pu, Yuanyuan
    Zhao, Zhengpeng
    Nie, Rencan
    Xu, Dan
    Qian, Wenhua
    Wu, Hao
    [J]. NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2103 - 2125
  • [37] Multimodal sentiment analysis using hierarchical fusion with context modeling
    Majumder, N.
    Hazarika, D.
    Gelbukh, A.
    Cambria, E.
    Poria, S.
    [J]. KNOWLEDGE-BASED SYSTEMS, 2018, 161 : 124 - 133
  • [38] TCHFN: Multimodal sentiment analysis based on Text-Centric Hierarchical Fusion Network
    Hou, Jingming
    Omar, Nazlia
    Tiun, Sabrina
    Saad, Saidah
    He, Qian
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [39] Text-centered cross-sample fusion network for multimodal sentiment analysis
    Huang, Qionghao
    Chen, Jili
    Huang, Changqin
    Huang, Xiaodi
    Wang, Yi
    [J]. MULTIMEDIA SYSTEMS, 2024, 30 (04)
  • [40] TeFNA: Text-centered fusion network with crossmodal attention for multimodal sentiment analysis
    Huang, Changqin
    Zhang, Junling
    Wu, Xuemei
    Wang, Yi
    Li, Ming
    Huang, Xiaodi
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 269