Modeling inter-modal incongruous sentiment expressions for multi-modal sarcasm detection

被引:2
|
作者
Ou, Lisong [1 ,2 ,3 ]
Li, Zhixin [1 ,2 ]
机构
[1] Guangxi Normal Univ, Key Lab Educ Blockchain & Intelligent Technol, Minist Educ, Guilin 541004, Peoples R China
[2] Guangxi Normal Univ, Guangxi Key Lab Multisource Informat Min & Secur, Guilin 541004, Peoples R China
[3] Guilin Univ Technol, Sch Math & Stat, Guilin 541004, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-modal sarcasm detection; Graph convolutional network; Cross-modal mapping; External knowledge; Cross-correlation graphs;
D O I
10.1016/j.neucom.2024.128874
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modal sarcasm detection (MSD) presents a formidable and intricate endeavor. Despite strides made by extant models, two principal hurdles persist: Firstly, prevailing methodologies merely address superficial disparities between textual inputs and associated images, neglecting nuanced inter-modal combinations. Secondly, satirical instances frequently involve intricate emotional expressions, highlighting the imperative of leveraging emotional cues across modalities to discern satirical nuances. Accordingly, this research proposes the utilization of a deep graph convolutional network that integrates cross-modal mapping information to effectively identify significant incongruent sentiment expressions across various modalities for the purpose of multi-modal sarcasm detection. Specifically, we first design a cross-modal mapping network, which obtains the interaction information between these two modalities by mapping text feature vectors and image feature vectors two by two to compensate for the lack of multi-modal data in the fusion process. Additionally, we employ external knowledge of ANPS as abridge to construct cross-correlation graphs through highly correlated satirical cues and their connection weights between image and text modalities. Afterward, the GCN architecture with retrieval-based attentional mechanisms will effectively capture satirical cues. The experiments conducted on two publicly available datasets demonstrate a significant enhancement in the performance of our method when compared to numerous contemporary models.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Inter-modal attention: ERPs to auditory targets in an inter-modal oddball task
    Brown, Christopher R.
    Clarke, Adam R.
    Barry, Robert J.
    INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2006, 62 (01) : 77 - 86
  • [22] Multi-modal sarcasm detection based on Multi-Channel Enhanced Fusion model
    Fang, Hong
    Liang, Dahao
    Xiang, Weiyu
    NEUROCOMPUTING, 2024, 578
  • [23] Self-Adaptive Representation Learning Model for Multi-Modal Sentiment and Sarcasm Joint Analysis
    Zhang, Yazhou
    Yu, Yang
    Wang, Mengyao
    Huang, Min
    Hossain, M. Shamim
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (05)
  • [24] Multi-modal Synthesis of Regular Expressions
    Chen, Qiaochu
    Wang, Xinyu
    Ye, Xi
    Durrett, Greg
    Dillig, Isil
    PROCEEDINGS OF THE 41ST ACM SIGPLAN CONFERENCE ON PROGRAMMING LANGUAGE DESIGN AND IMPLEMENTATION (PLDI '20), 2020, : 487 - 502
  • [25] MV-BART: Multi-view BART for Multi-modal Sarcasm Detection
    Zhuang, Xingjie
    Zhou, Fengling
    Li, Zhixin
    PROCEEDINGS OF THE 33RD ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2024, 2024, : 3602 - 3611
  • [26] Multi-Modal Sarcasm Detection and Humor Classification in Code-Mixed Conversations
    Bedi, Manjot
    Kumar, Shivani
    Akhtar, Md Shad
    Chakraborty, Tanmoy
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (02) : 1363 - 1375
  • [27] MMSD2.0: Towards a Reliable Multi-modal Sarcasm Detection System
    Qin, Libo
    Huang, Shijue
    Chen, Qiguang
    Cai, Chenran
    Zhang, Yudi
    Bin Liang
    Che, Wanxiang
    Xu, Ruifeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10834 - 10845
  • [28] Single-Stage Extensive Semantic Fusion for multi-modal sarcasm detection
    Fang, Hong
    Liang, Dahao
    Xiang, Weiyu
    ARRAY, 2024, 22
  • [29] Mutual-Enhanced Incongruity Learning Network for Multi-Modal Sarcasm Detection
    Qiao, Yang
    Jing, Liqiang
    Song, Xuemeng
    Chen, Xiaolin
    Zhu, Lei
    Nie, Liqiang
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9507 - 9515
  • [30] A Multitask Framework for Sentiment, Emotion and Sarcasm aware Cyberbullying Detection from Multi-modal Code-Mixed Memes
    Maity, Krishanu
    Jha, Prince
    Saha, Sriparna
    Bhattacharyya, Pushpak
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1739 - 1749