Multi-modal sarcasm detection based on Multi-Channel Enhanced Fusion model

被引:7
|
作者
Fang, Hong [1 ]
Liang, Dahao [2 ]
Xiang, Weiyu [2 ]
机构
[1] Shanghai Polytech Univ, Sch Math Phys & Stat, Shanghai 201209, Peoples R China
[2] Shanghai Polytech Univ, Inst Artificial Intelligence, Sch Comp & Informat Engn, Shanghai 201209, Peoples R China
关键词
Multi-modal sarcasm detection; Attention mechanism; Feature fusion;
D O I
10.1016/j.neucom.2024.127440
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The voluminous quantity of data accessible on social media platforms offers insight into the sentiment disposition of individual users, where multi -modal sarcasm detection is often confounding. Existing sarcasm detection methods use different information fusion methods to combine information from different modalities but ignore hidden information within modalities and inconsistent information between modalities. Discovering the implicit information within the modalities and strengthening the information interaction between modalities is still an important challenge. In this paper, we propose a Multi -Channel Enhanced Fusion (MCEF) model for cross -modal sarcasm detection to maximize the information extraction between different modalities. Specifically, text extracted from images acts as a new modality in the front-end fusion models to augment the utilization of image semantic information. Then, we propose a novel bipolar semantic attention mechanism to uncover the inconsistencies among different modal features. Furthermore, a decision -level fusion strategy from a new perspective is devised based on four models to achieve multi -channel fusion, each with a distinct focus, to leverage their advantages and mitigate the limitations. Extensive experiments demonstrate that our model surpasses current state-of-the-art models in multi -modal sarcasm detection.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] A multi-modal fusion YoLo network for traffic detection
    Zheng, Xinwang
    Zheng, Wenjie
    Xu, Chujie
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (02)
  • [42] Fake News Detection Based on BERT Multi-domain and Multi-modal Fusion Network
    Yu, Kai
    Jiao, Shiming
    Ma, Zhilong
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 252
  • [43] RDMIF: Reverse dictionary model based on multi-modal information fusion
    Tian, Sicheng
    Huang, Shaobin
    Li, Rongsheng
    Wei, Chi
    NEUROCOMPUTING, 2025, 619
  • [44] Multi-modal pedestrian detection with misalignment based on modal-wise regression and multi-modal IoU
    Wanchaitanawong, Napat
    Tanaka, Masayuki
    Shibata, Takashi
    Okutomi, Masatoshi
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (01)
  • [45] Multi-modal Intermediate Fusion Model for diagnosis prediction
    Lu, You
    Niu, Ke
    Peng, Xueping
    Zeng, Jingni
    Pei, Su
    6TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE, ICIAI2022, 2022, : 38 - 43
  • [46] Multi-Modal AI for Enhanced Forest Fire Early Detection: Scalar and Image Fusion
    Azzabi, Tarek
    Jeridi, Mohamed Hechmi
    Mejri, Ikbel
    Ezzedine, Tahar
    2024 IEEE INTERNATIONAL CONFERENCE ON ADVANCED SYSTEMS AND EMERGENT TECHNOLOGIES, ICASET 2024, 2024,
  • [47] Multi-Modal Sensor Fusion and Selection for Enhanced Situational Awareness
    Reily, Brian
    Reardon, Christopher
    Zhang, Hao
    VIRTUAL, AUGMENTED, AND MIXED REALITY (XR) TECHNOLOGY FOR MULTI-DOMAIN OPERATIONS II, 2021, 11759
  • [48] Multi-Modal Fusion for Multi-Task Fuzzy Detection of Rail Anomalies
    Liyuan, Yang
    Osman, Ghazali
    Abdul Rahman, Safawi
    Mustapha, Muhammad Firdaus
    IEEE ACCESS, 2024, 12 : 73925 - 73935
  • [49] INTELLIGENT ADVANCED ATTACK DETECTION TECHNOLOGY BASED ON MULTI-MODAL DATA FUSION
    HANG F.
    XIE L.
    ZHANG Z.
    HU J.I.A.N.
    Scalable Computing, 2024, 25 (04): : 2581 - 2588
  • [50] Cardiovascular disease detection based on deep learning and multi-modal data fusion
    Zhu, Jiayuan
    Liu, Hui
    Liu, Xiaowei
    Chen, Chao
    Shu, Minglei
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 99