Reasoning with Multimodal Sarcastic Tweets via Modeling Cross-Modality Contrast and Semantic Association

被引:0
|
作者
Xu, Nan [1 ]
Zeng, Zhixiong [1 ]
Mao, Wenji [1 ]
机构
[1] Univ Chinese Acad Sci, Chinese Acad Sci, Inst Automat, Sch Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sarcasm is a sophisticated linguistic phenomenon to express the opposite of what one really means. With the rapid growth of social media, multimodal sarcastic tweets are widely posted on various social platforms. In multimodal context, sarcasm is no longer a pure linguistic phenomenon, and due to the nature of social media short text, the opposite is more often manifested via cross-modality expressions. Thus traditional text-based methods are insufficient to detect multimodal sarcasm. To reason with multimodal sarcastic tweets, in this paper, we propose a novel method for modeling cross-modality contrast in the associated context. Our method models both cross-modality contrast and semantic association by constructing the Decomposition and Relation Network (namely D&R Net). The decomposition network represents the commonality and discrepancy between image and text, and the relation network models the semantic association in cross-modality context. Experimental results on a public dataset demonstrate the effectiveness of our model in multimodal sarcasm detection.
引用
收藏
页码:3777 / 3786
页数:10
相关论文
共 50 条
  • [1] Hierarchical Cross-Modality Semantic Correlation Learning Model for Multimodal Summarization
    Zhang, Litian
    Zhang, Xiaoming
    Pan, Junshu
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11676 - 11684
  • [2] CISum: Learning Cross-modality Interaction to Enhance Multimodal Semantic Coverage for Multimodal Summarization
    Zhang, Litian
    Zhang, Xiaoming
    Guo, Ziming
    Liu, Zhipeng
    [J]. PROCEEDINGS OF THE 2023 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2023, : 370 - 378
  • [3] Cross-Modality Learning by Exploring Modality Interactions for Emotion Reasoning
    Tran, Thi-Dung
    Ho, Ngoc-Huynh
    Pant, Sudarshan
    Yang, Hyung-Jeong
    Kim, Soo-Hyung
    Lee, Gueesang
    [J]. IEEE ACCESS, 2023, 11 : 56634 - 56648
  • [4] Cross-Modality Semantic Integration With Hypothesis Rescoring for Robust Interpretation of Multimodal User Interactions
    Hui, Pui-Yu
    Meng, Helen M.
    [J]. IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2009, 17 (03): : 486 - 500
  • [5] Scaling Multimodal Pre-Training via Cross-Modality Gradient Harmonization
    Wu, Junru
    Liang, Yi
    Han, Feng
    Akbari, Hassan
    Wang, Zhangyang
    Yu, Cong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [6] iSCAN: Automatic Speaker Adaptation via Iterative Cross-modality Association
    Xiangli, Yuanbo
    Lu, Chris Xiaoxuan
    Zhao, Peijun
    Chen, Changhao
    Markham, Andrew
    [J]. UBICOMP/ISWC'19 ADJUNCT: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING AND PROCEEDINGS OF THE 2019 ACM INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS, 2019, : 529 - 533
  • [7] CMOT: Cross-Modality Optimal Transport for multimodal inference
    Sayali Anil Alatkar
    Daifeng Wang
    [J]. Genome Biology, 24
  • [8] CMOT: Cross-Modality Optimal Transport for multimodal inference
    Alatkar, Sayali Anil
    Wang, Daifeng
    [J]. GENOME BIOLOGY, 2023, 24 (01)
  • [9] SENTENCE AND PICTURE MEMORY - CROSS-MODALITY SEMANTIC INTEGRATION
    PEZDEK, K
    MARSH, G
    [J]. BULLETIN OF THE PSYCHONOMIC SOCIETY, 1975, 6 (NB4) : 435 - 435
  • [10] CROSS-MODALITY SEMANTIC INTEGRATION OF SENTENCE AND PICTURE MEMORY
    PEZDEK, K
    [J]. JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN LEARNING AND MEMORY, 1977, 3 (05): : 515 - 524