Multi-modal sarcasm detection based on Multi-Channel Enhanced Fusion model

被引:7
|
作者
Fang, Hong [1 ]
Liang, Dahao [2 ]
Xiang, Weiyu [2 ]
机构
[1] Shanghai Polytech Univ, Sch Math Phys & Stat, Shanghai 201209, Peoples R China
[2] Shanghai Polytech Univ, Inst Artificial Intelligence, Sch Comp & Informat Engn, Shanghai 201209, Peoples R China
关键词
Multi-modal sarcasm detection; Attention mechanism; Feature fusion;
D O I
10.1016/j.neucom.2024.127440
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The voluminous quantity of data accessible on social media platforms offers insight into the sentiment disposition of individual users, where multi -modal sarcasm detection is often confounding. Existing sarcasm detection methods use different information fusion methods to combine information from different modalities but ignore hidden information within modalities and inconsistent information between modalities. Discovering the implicit information within the modalities and strengthening the information interaction between modalities is still an important challenge. In this paper, we propose a Multi -Channel Enhanced Fusion (MCEF) model for cross -modal sarcasm detection to maximize the information extraction between different modalities. Specifically, text extracted from images acts as a new modality in the front-end fusion models to augment the utilization of image semantic information. Then, we propose a novel bipolar semantic attention mechanism to uncover the inconsistencies among different modal features. Furthermore, a decision -level fusion strategy from a new perspective is devised based on four models to achieve multi -channel fusion, each with a distinct focus, to leverage their advantages and mitigate the limitations. Extensive experiments demonstrate that our model surpasses current state-of-the-art models in multi -modal sarcasm detection.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Multi-Modal Sarcasm Detection in Twitter with Hierarchical Fusion Model
    Cai, Yitao
    Cai, Huiyu
    Wan, Xiaojun
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2506 - 2515
  • [2] Attention-based multi-modal fusion sarcasm detection
    Liu, Jing
    Tian, Shengwei
    Yu, Long
    Long, Jun
    Zhou, Tiejun
    Wang, Bo
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2097 - 2108
  • [3] A multi-modal sarcasm detection model based on cue learning
    Lu, Ming
    Dong, Zhiqiang
    Guo, Ziming
    Zhang, Xiaoming
    Lu, Xinxi
    Wang, Tianbo
    Zhang, Litian
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [4] Multi-modal sarcasm detection using ensemble net model
    Sukhavasi, Vidyullatha
    Dondeti, Venkatesulu
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (01) : 403 - 425
  • [5] Multi-Modal Sarcasm Detection Based on Dual Generative Processes
    Ma, Huiying
    He, Dongxiao
    Wang, Xiaobao
    Jin, Di
    Ge, Meng
    Wang, Longbiao
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 2279 - 2287
  • [6] Multi-modal Sarcasm Detection on Social Media via Multi-Granularity Information Fusion
    Ou, Lisong
    Li, Zhixin
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2025, 21 (03)
  • [7] Single-Stage Extensive Semantic Fusion for multi-modal sarcasm detection
    Fang, Hong
    Liang, Dahao
    Xiang, Weiyu
    ARRAY, 2024, 22
  • [8] Multi-Modal Multi-Channel Target Speech Separation
    Gu, Rongzhi
    Zhang, Shi-Xiong
    Xu, Yong
    Chen, Lianwu
    Zou, Yuexian
    Yu, Dong
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (03) : 530 - 541
  • [9] Multi-Modal Sarcasm Detection with Sentiment Word Embedding
    Fu, Hao
    Liu, Hao
    Wang, Hongling
    Xu, Linyan
    Lin, Jiali
    Jiang, Dazhi
    ELECTRONICS, 2024, 13 (05)
  • [10] Mutual-Enhanced Incongruity Learning Network for Multi-Modal Sarcasm Detection
    Qiao, Yang
    Jing, Liqiang
    Song, Xuemeng
    Chen, Xiaolin
    Zhu, Lei
    Nie, Liqiang
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9507 - 9515