Modeling inter-modal incongruous sentiment expressions for multi-modal sarcasm detection

被引:2
|
作者
Ou, Lisong [1 ,2 ,3 ]
Li, Zhixin [1 ,2 ]
机构
[1] Guangxi Normal Univ, Key Lab Educ Blockchain & Intelligent Technol, Minist Educ, Guilin 541004, Peoples R China
[2] Guangxi Normal Univ, Guangxi Key Lab Multisource Informat Min & Secur, Guilin 541004, Peoples R China
[3] Guilin Univ Technol, Sch Math & Stat, Guilin 541004, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-modal sarcasm detection; Graph convolutional network; Cross-modal mapping; External knowledge; Cross-correlation graphs;
D O I
10.1016/j.neucom.2024.128874
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modal sarcasm detection (MSD) presents a formidable and intricate endeavor. Despite strides made by extant models, two principal hurdles persist: Firstly, prevailing methodologies merely address superficial disparities between textual inputs and associated images, neglecting nuanced inter-modal combinations. Secondly, satirical instances frequently involve intricate emotional expressions, highlighting the imperative of leveraging emotional cues across modalities to discern satirical nuances. Accordingly, this research proposes the utilization of a deep graph convolutional network that integrates cross-modal mapping information to effectively identify significant incongruent sentiment expressions across various modalities for the purpose of multi-modal sarcasm detection. Specifically, we first design a cross-modal mapping network, which obtains the interaction information between these two modalities by mapping text feature vectors and image feature vectors two by two to compensate for the lack of multi-modal data in the fusion process. Additionally, we employ external knowledge of ANPS as abridge to construct cross-correlation graphs through highly correlated satirical cues and their connection weights between image and text modalities. Afterward, the GCN architecture with retrieval-based attentional mechanisms will effectively capture satirical cues. The experiments conducted on two publicly available datasets demonstrate a significant enhancement in the performance of our method when compared to numerous contemporary models.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Contextual Inter-modal Attention for Multi-modal Sentiment Analysis
    Ghosal, Deepanway
    Akhtar, Md Shad
    Chauhan, Dushyant
    Poria, Soujanya
    Ekbalt, Asif
    Bhattacharyyat, Pushpak
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 3454 - 3466
  • [2] Multi-Modal Sarcasm Detection with Sentiment Word Embedding
    Fu, Hao
    Liu, Hao
    Wang, Hongling
    Xu, Linyan
    Lin, Jiali
    Jiang, Dazhi
    ELECTRONICS, 2024, 13 (05)
  • [3] Modeling Intra and Inter-modality Incongruity for Multi-Modal Sarcasm Detection
    Pan, Hongliang
    Lin, Zheng
    Fu, Peng
    Qi, Yatao
    Wang, Weiping
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1383 - 1392
  • [4] A Multi-Modal Entity Alignment Method with Inter-Modal Enhancement
    Yuan, Song
    Lu, Zexin
    Li, Qiyuan
    Gu, Jinguang
    BIG DATA AND COGNITIVE COMPUTING, 2023, 7 (02)
  • [5] Multi-Modal Sarcasm Detection with Interactive In-Modal and Cross-Modal Graphs
    Liang, Bin
    Lou, Chenwei
    Li, Xiang
    Gui, Lin
    Yang, Min
    Xu, Ruifeng
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4707 - 4715
  • [6] Learning Inter-Modal Correspondence and Phenotypes From Multi-Modal Electronic Health Records
    Yin, Kejing
    Cheung, William K.
    Fung, Benjamin C. M.
    Poon, Jonathan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (09) : 4328 - 4341
  • [7] Cross-modal incongruity aligning and collaborating for multi-modal sarcasm detection
    Wang, Jie
    Yang, Yan
    Jiang, Yongquan
    Ma, Minbo
    Xie, Zhuyang
    Li, Tianrui
    INFORMATION FUSION, 2024, 103
  • [8] Sentiment and Emotion help Sarcasm? A Multi-task Learning Framework for Multi-Modal Sarcasm, Sentiment and Emotion Analysis
    Chauhan, Dushyant Singh
    Dhanush, S. R.
    Ekbal, Asif
    Bhattacharyya, Pushpak
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 4351 - 4360
  • [9] IIFDD: Intra and inter-modal fusion for depression detection with multi-modal information from Internet of Medical Things
    Chen, Jian
    Hu, Yuzhu
    Lai, Qifeng
    Wang, Wei
    Chen, Junxin
    Liu, Han
    Srivastava, Gautam
    Bashir, Ali Kashif
    Hu, Xiping
    INFORMATION FUSION, 2024, 102
  • [10] Multi-Modal Sarcasm Detection via Cross-Modal Graph Convolutional Network
    Liang, Bin
    Lou, Chenwei
    Li, Xiang
    Yang, Min
    Gui, Lin
    He, Yulan
    Pei, Wenjie
    Xu, Ruifeng
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1767 - 1777