Multi-Model Fusion Framework Using Deep Learning for Visual-Textual Sentiment Classification

被引:0
|
作者
Salman Al-Tameemi I.K. [1 ,2 ,3 ]
Feizi-Derakhshi M.-R. [1 ,2 ]
Pashazadeh S. [2 ]
Asadpour M. [2 ]
机构
[1] Computerized Intelligence Systems Laboratory, Department of Computer Engineering, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz
[2] Department of Computer Engineering, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz
[3] State Company for Engineering Rehabilitation and Testing, Iraqi Ministry of Industry and Minerals, Baghdad
来源
Computers, Materials and Continua | 2023年 / 76卷 / 02期
关键词
decision fusion; deep learning; interpretability; joint fusion; multimodal classification; Sentiment analysis;
D O I
10.32604/CMC.2023.040997
中图分类号
学科分类号
摘要
Multimodal Sentiment Analysis (SA) is gaining popularity due to its broad application potential. The existing studies have focused on the SA of single modalities, such as texts or photos, posing challenges in effectively handling social media data with multiple modalities. Moreover, most multimodal research has concentrated on merely combining the two modalities rather than exploring their complex correlations, leading to unsatisfactory sentiment classification results. Motivated by this, we propose a new visual-textual sentiment classification model named Multi-Model Fusion (MMF), which uses a mixed fusion framework for SA to effectively capture the essential information and the intrinsic relationship between the visual and textual content. The proposed model comprises three deep neural networks. Two different neural networks are proposed to extract the most emotionally relevant aspects of image and text data. Thus, more discriminative features are gathered for accurate sentiment classification. Then, a multichannel joint fusion model with a self-attention technique is proposed to exploit the intrinsic correlation between visual and textual characteristics and obtain emotionally rich information for joint sentiment classification. Finally, the results of the three classifiers are integrated using a decision fusion scheme to improve the robustness and generalizability of the proposed model. An interpretable visual-textual sentiment classification model is further developed using the Local Interpretable Model-agnostic Explanation model (LIME) to ensure the model’s explainability and resilience. The proposed MMF model has been tested on four real-world sentiment datasets, achieving (99.78%) accuracy on Binary_Getty (BG), (99.12%) on Binary_iStock (BIS), (95.70%) on Twitter, and (79.06%) on the Multi-View Sentiment Analysis (MVSA) dataset. These results demonstrate the superior performance of our MMF model compared to single-model approaches and current state-of-the-art techniques based on model evaluation criteria. © 2023 Tech Science Press. All rights reserved.
引用
收藏
页码:2145 / 2177
页数:32
相关论文
共 50 条
  • [31] Sentiment based emotion classification in unstructured textual data using dual stage deep model
    Padminivalli, V. S. J. R. K.
    Rao, M. V. P. Chandra Sekhara
    Narne, Naga Sai Ram
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (8) : 22875 - 22907
  • [32] Sentiment based emotion classification in unstructured textual data using dual stage deep model
    S J R K Padminivalli V
    M. V. P. Chandra Sekhara Rao
    Naga Sai Ram Narne
    Multimedia Tools and Applications, 2024, 83 : 22875 - 22907
  • [33] Spatially Attentive Visual Tracking Using Multi-Model Adaptive Response Fusion
    Zhang, Jianming
    Wu, You
    Feng, Wenjun
    Wang, Jin
    IEEE ACCESS, 2019, 7 : 83873 - 83887
  • [34] Multi-Class Document Image Classification using Deep Visual and Textual Features
    Sevim, Semih
    Ekinci, Ekin
    Omurca, Sevinc Ilhan
    Edinc, Eren Berk
    Eken, Suleyman
    Erdem, Turkucan
    Sayar, Ahmet
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2022, 21 (02)
  • [35] Texture classification using multi-model feature integration by Bayesian fusion.
    Huang, Y
    Chan, KL
    Huang, Z
    Wang, L
    2001 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-VI, PROCEEDINGS: VOL I: SPEECH PROCESSING 1; VOL II: SPEECH PROCESSING 2 IND TECHNOL TRACK DESIGN & IMPLEMENTATION OF SIGNAL PROCESSING SYSTEMS NEURALNETWORKS FOR SIGNAL PROCESSING; VOL III: IMAGE & MULTIDIMENSIONAL SIGNAL PROCESSING MULTIMEDIA SIGNAL PROCESSING - VOL IV: SIGNAL PROCESSING FOR COMMUNICATIONS; VOL V: SIGNAL PROCESSING EDUCATION SENSOR ARRAY & MULTICHANNEL SIGNAL PROCESSING AUDIO & ELECTROACOUSTICS; VOL VI: SIGNAL PROCESSING THEORY & METHODS STUDENT FORUM, 2001, : 4052 - 4052
  • [36] Deep Multi-Model Fusion for Human Activity Recognition Using Evolutionary Algorithms
    Verma, Kamal Kant
    Singh, Brij Mohan
    INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2021, 7 (02): : 44 - 58
  • [37] Visual audio and textual triplet fusion network for multi-modal sentiment analysis
    Lv, Cai-Chao
    Zhang, Xuan
    Zhang, Hong-Bo
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (12) : 9505 - 9513
  • [38] A deep learning framework for sarcastic sentiment classification in opinion polls
    Fredrick Boafo
    Solomon Mensah
    Elizabeth Akuafum Dick
    Journal of Data, Information and Management, 2023, 5 (4): : 333 - 343
  • [39] BAITRADAR: A MULTI-MODEL CLICKBAIT DETECTION ALGORITHM USING DEEP LEARNING
    Gamage, Bhanuka
    Labib, Adnan
    Joomun, Aisha
    Lim, Chern Hong
    Wong, KokSheik
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2665 - 2669
  • [40] An Ensemble of Deep Learning-Based Multi-Model for ECG Heartbeats Arrhythmia Classification
    Essa, Ehab
    Xie, Xianghua
    IEEE Access, 2021, 9 : 103452 - 103464