Multi-modal sarcasm detection based on Multi-Channel Enhanced Fusion model

被引:7
|
作者
Fang, Hong [1 ]
Liang, Dahao [2 ]
Xiang, Weiyu [2 ]
机构
[1] Shanghai Polytech Univ, Sch Math Phys & Stat, Shanghai 201209, Peoples R China
[2] Shanghai Polytech Univ, Inst Artificial Intelligence, Sch Comp & Informat Engn, Shanghai 201209, Peoples R China
关键词
Multi-modal sarcasm detection; Attention mechanism; Feature fusion;
D O I
10.1016/j.neucom.2024.127440
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The voluminous quantity of data accessible on social media platforms offers insight into the sentiment disposition of individual users, where multi -modal sarcasm detection is often confounding. Existing sarcasm detection methods use different information fusion methods to combine information from different modalities but ignore hidden information within modalities and inconsistent information between modalities. Discovering the implicit information within the modalities and strengthening the information interaction between modalities is still an important challenge. In this paper, we propose a Multi -Channel Enhanced Fusion (MCEF) model for cross -modal sarcasm detection to maximize the information extraction between different modalities. Specifically, text extracted from images acts as a new modality in the front-end fusion models to augment the utilization of image semantic information. Then, we propose a novel bipolar semantic attention mechanism to uncover the inconsistencies among different modal features. Furthermore, a decision -level fusion strategy from a new perspective is devised based on four models to achieve multi -channel fusion, each with a distinct focus, to leverage their advantages and mitigate the limitations. Extensive experiments demonstrate that our model surpasses current state-of-the-art models in multi -modal sarcasm detection.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Multi-modal deep fusion based fake news detection method
    Jing Q.
    Fan X.
    Wang B.
    Bi J.
    Tan H.
    High Technology Letters, 2022, 32 (04) : 392 - 403
  • [32] Citrus Huanglongbing Detection Based on Multi-Modal Feature Fusion Learning
    Yang, Dongzi
    Wang, Fengcheng
    Hu, Yuqi
    Lan, Yubin
    Deng, Xiaoling
    FRONTIERS IN PLANT SCIENCE, 2021, 12
  • [33] Lightweight video salient object detection via channel-shuffle enhanced multi-modal fusion network
    Huang, Kan
    Xu, Zhijing
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (1) : 1025 - 1039
  • [34] Multi-task & Multi-modal Sentiment Analysis Model Based on Aware Fusion
    Wu S.
    Ma J.
    Data Analysis and Knowledge Discovery, 2023, 7 (10) : 74 - 84
  • [35] An Abnormal External Link Detection Algorithm Based on Multi-Modal Fusion
    Wu, Zhiqiang
    INTERNATIONAL JOURNAL OF INFORMATION SECURITY AND PRIVACY, 2024, 18 (01)
  • [36] Lightweight video salient object detection via channel-shuffle enhanced multi-modal fusion network
    Kan Huang
    Zhijing Xu
    Multimedia Tools and Applications, 2024, 83 : 1025 - 1039
  • [37] MULTI-MODAL FUSION ENHANCED MODEL FOR DRIVER'S FACIAL EXPRESSION RECOGNITION
    Chen, Jianrong
    De, Sujit
    Wang, Lei
    Bi, Ning
    Liu, Peng
    2021 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2021,
  • [38] MMFusion: A Generalized Multi-Modal Fusion Detection Framework
    Cui, Leichao
    Li, Xiuxian
    Meng, Min
    Mo, Xiaoyu
    2023 IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING, ICDL, 2023, : 415 - 422
  • [39] Improving multi-modal data fusion by anomaly detection
    Jakub Simanek
    Vladimir Kubelka
    Michal Reinstein
    Autonomous Robots, 2015, 39 : 139 - 154
  • [40] Improving multi-modal data fusion by anomaly detection
    Simanek, Jakub
    Kubelka, Vladimir
    Reinstein, Michal
    AUTONOMOUS ROBOTS, 2015, 39 (02) : 139 - 154