MIAN: Multi-head Incongruity Aware Attention Network with transfer learning for sarcasm detection

被引:0
|
作者
Guan, Xin [1 ]
Cao, Jiuxin [1 ]
Zhang, Hui [1 ]
Cao, Biwei [1 ]
Liu, Bo [2 ]
机构
[1] Southeast Univ, Sch Cyber Sci & Engn, Nanjing 211189, Peoples R China
[2] Southeast Univ, Sch Comp Sci & Engn, Nanjing 211189, Peoples R China
基金
中国国家自然科学基金;
关键词
Sentiment analysis; Sarcasm detection; Transfer learning; Deep neural networks; SENTIMENT;
D O I
10.1016/j.eswa.2024.125702
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sarcasm is a common rhetorical metaphor in social media platforms, that individuals express emotion contrary to the literal meaning. Capturing the incongruity in the texts is the critical factor in sarcasm detection. Although several studies have looked at the incongruity of a single text, there is currently a lack of studies on modeling the incongruity of contextual information. Inspired by Multi-Head Attention mechanism from Transformer, we propose a Multi-head Incongruity Aware Attention Network, which concentrates on both target semantic incongruity and contextual semantic incongruity. Specifically, we design a multi-head self-match network to capture target semantic incongruity in a single text. Moreover, a multi-head co-match network is applied to model the contextual semantic incongruity. Furthermore, due to the scarcity of sarcasm data and considering the correlation between tasks of sentiment analysis and sarcasm detection, we pre-train the language model with a great amount of sentiment analysis data, which enhances its ability to capture sentimental features in the text. The experimental results demonstrate that our model achieves state-of-the-art performance on four benchmark datasets, with an accuracy gain of 3.8% on Tweets Ghost, 1.1% on SARC Pol, 1.9% on Ciron and an F1-Score gain of 0.3% on FigLang Twitter.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Bilinear Multi-Head Attention Graph Neural Network for Traffic Prediction
    Hu, Haibing
    Han, Kai
    Yin, Zhizhuo
    ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, 2022, : 33 - 43
  • [42] Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis
    Zhang, Rui
    Xue, Chengrong
    Qi, Qingfu
    Lin, Liyuan
    Zhang, Jing
    Zhang, Lun
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [43] A Dual Multi-Head Contextual Attention Network for Hyperspectral Image Classification
    Liang, Miaomiao
    He, Qinghua
    Yu, Xiangchun
    Wang, Huai
    Meng, Zhe
    Jiao, Licheng
    REMOTE SENSING, 2022, 14 (13)
  • [44] Acoustic Word Embedding Based on Multi-Head Attention Quadruplet Network
    Zhu, Shirong
    Zhang, Ying
    He, Kai
    Zhao, Lasheng
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 184 - 188
  • [45] Siamese Network cooperating with Multi-head Attention for semantic sentence matching
    Yuan, Zhao
    Jun, Sun
    2020 19TH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED COMPUTING AND APPLICATIONS FOR BUSINESS ENGINEERING AND SCIENCE (DCABES 2020), 2020, : 235 - 238
  • [46] Personalized multi-head self-attention network for news recommendation
    Zheng, Cong
    Song, Yixuan
    NEURAL NETWORKS, 2025, 181
  • [47] Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition
    Wen, Zhengyao
    Lin, Wenzhong
    Wang, Tao
    Xu, Ge
    BIOMIMETICS, 2023, 8 (02)
  • [48] Fact-sentiment incongruity combination network for multimodal sarcasm detection
    Lu, Qiang
    Long, Yunfei
    Sun, Xia
    Feng, Jun
    Zhang, Hao
    INFORMATION FUSION, 2024, 104
  • [49] Capsule Network Improved Multi-Head Attention for Word Sense Disambiguation
    Cheng, Jinfeng
    Tong, Weiqin
    Yan, Weian
    APPLIED SCIENCES-BASEL, 2021, 11 (06):
  • [50] Multi-Head Attention with Diversity for Learning Grounded Multilingual Multimodal Representations
    Huang, Po-Yao
    Chang, Xiaojun
    Hauptmann, Alexander
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1461 - 1467