Multimodal Sentiment Classifier Framework for Different Scene Contexts

被引:0
|
作者
Silva, Nelson [1 ]
Cardoso, Pedro J. S. [1 ]
Rodrigues, Joao M. F. [1 ]
机构
[1] Univ Algarve, NOVA LINCS & ISE, P-8005139 Faro, Portugal
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 16期
关键词
sentiment analysis; affective computing; human-centered AI; Multimodal Sentiment Classifier; FACIAL EXPRESSION RECOGNITION; EMOTION;
D O I
10.3390/app14167065
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Sentiment analysis (SA) is an effective method for determining public opinion. Social media posts have been the subject of much research, due to the platforms' enormous and diversified user bases that regularly share thoughts on nearly any subject. However, on posts composed by a text-image pair, the written description may or may not convey the same sentiment as the image. The present study uses machine learning models for the automatic sentiment evaluation of pairs of text and image(s). The sentiments derived from the image and text are evaluated independently and merged (or not) to form the overall sentiment, returning the sentiment of the post and the discrepancy between the sentiments represented by the text-image pair. The image sentiment classification is divided into four categories-"indoor" (IND), "man-made outdoors" (OMM), "non-man-made outdoors" (ONMM), and "indoor/outdoor with persons in the background" (IOwPB)-and then ensembled into an image sentiment classification model (ISC), that can be compared with a holistic image sentiment classifier (HISC), showing that the ISC achieves better results than the HISC. For the Flickr sub-data set, the sentiment classification of images achieved an accuracy of 68.50% for IND, 83.20% for OMM, 84.50% for ONMM, 84.80% for IOwPB, and 76.45% for ISC, compared to 65.97% for the HISC. For the text sentiment classification, in a sub-data set of B-T4SA, an accuracy of 92.10% was achieved. Finally, the text-image combination, in the authors' private data set, achieved an accuracy of 78.84%.
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Multimodal Mutual Attention-Based Sentiment Analysis Framework Adapted to Complicated Contexts
    He, Lijun
    Wang, Ziqing
    Wang, Liejun
    Li, Fan
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7131 - 7143
  • [2] A Multitask Learning Framework for Multimodal Sentiment Analysis
    Jiang, Dazhi
    Wei, Runguo
    Liu, Hao
    Wen, Jintao
    Tu, Geng
    Zheng, Lin
    Cambria, Erik
    [J]. 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 151 - 157
  • [3] A quantum-inspired multimodal sentiment analysis framework
    Zhang, Yazhou
    Song, Dawei
    Zhang, Peng
    Wang, Panpan
    Li, Jingfei
    Li, Xiang
    Wang, Benyou
    [J]. THEORETICAL COMPUTER SCIENCE, 2018, 752 : 21 - 40
  • [4] UniMF: A Unified Multimodal Framework for Multimodal Sentiment Analysis in Missing Modalities and Unaligned Multimodal Sequences
    Huan, Ruohong
    Zhong, Guowei
    Chen, Peng
    Liang, Ronghua
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 5753 - 5768
  • [5] Multimodal sentiment analysis using reliefF feature selection and random forest classifier
    Angadi S.
    Reddy V.S.
    [J]. International Journal of Computers and Applications, 2021, 43 (09) : 931 - 939
  • [6] Towards Multimodal Sentiment Analysis Inspired by the Quantum Theoretical Framework
    Li, Qiuchi
    Stefani, Alessandro
    Toto, Giovanni
    Di Buccio, Emanuele
    Melucci, Massimo
    [J]. THIRD INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2020), 2020, : 181 - 184
  • [7] Multi-token Fusion Framework for Multimodal Sentiment Analysis
    Long, Zhihui
    Deng, Huan
    Yang, Zhenguo
    Liu, Wenyin
    [J]. WEB AND BIG DATA, PT II, APWEB-WAIM 2023, 2024, 14332 : 424 - 438
  • [8] Multimodal sentiment analysis leveraging the strength of deep neural networks enhanced by the XGBoost classifier
    Chandrasekaran, Ganesh
    Dhanasekaran, S.
    Moorthy, C.
    Oli, A. Arul
    [J]. COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING, 2024,
  • [9] Accuracy of Multimodal Pedestrian Localisation in Different Terminal Contexts
    Riihijaervi, Janne
    Deng, Xiping
    Maehoenen, Petri
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS (PERCOM WORKSHOPS), 2018,
  • [10] A central multimodal fusion framework for outdoor scene image segmentation
    Zhang, Yifei
    Morel, Olivier
    Seulin, Ralph
    Meriaudeau, Fabrice
    Sidibe, Desire
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (09) : 12047 - 12060