Detecting Sarcasm in Conversation Context Using Transformer-Based Models

被引:0
|
作者
Avvaru, Adithya [1 ,2 ]
Vobilisetty, Sanath [2 ]
Mamidi, Radhika [1 ]
机构
[1] Int Inst Informat Technol, Hyderabad, India
[2] Teradata India Pvt Ltd, Mumbai, Maharashtra, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sarcasm detection, regarded as one of the subproblems of sentiment analysis, is a very typical task because the introduction of sarcastic words can flip the sentiment of the sentence itself. To date, many research works revolve around detecting sarcasm in one single sentence and there is very limited research to detect sarcasm resulting from multiple sentences. Current models used Long Short Term Memory (Hochreiter and Schmidhuber, 1997) (LSTM) variants with or without attention to detect sarcasm in conversations. We showed that the models using state-of-the-art Bidirectional Encoder Representations from Transformers (Devlin et al., 2018) (BERT), to capture syntactic and semantic information across conversation sentences, performed better than the current models. Based on the data analysis, we estimated that the number of sentences in the conversation that can contribute to the sarcasm and the results agrees to this estimation. We also perform a comparative study of our different versions of BERT-based model with other variants of LSTM model and XLNet (Yang et al, 2019) (both using the estimated number of conversation sentences) and find out that BERT-based models outperformed them.
引用
收藏
页码:98 / 103
页数:6
相关论文
共 50 条
  • [41] Transformer-Based Federated Learning Models for Recommendation Systems
    Reddy, M. Sujaykumar
    Karnati, Hemanth
    Sundari, L. Mohana
    IEEE ACCESS, 2024, 12 : 109596 - 109607
  • [42] A Comparison of Transformer-Based Language Models on NLP Benchmarks
    Greco, Candida Maria
    Tagarelli, Andrea
    Zumpano, Ester
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 490 - 501
  • [43] Context-free Transformer-based Generative Lemmatiser for Polish
    Walentynowicz, Wiktor
    Piasecki, Maciej
    Kot, Artur
    COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2022, 2022, 13501 : 207 - 219
  • [44] RadBERT: Adapting Transformer-based Language Models to Radiology
    Yan, An
    McAuley, Julian
    Lu, Xing
    Du, Jiang
    Chang, Eric Y.
    Gentili, Amilcare
    Hsu, Chun-Nan
    RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2022, 4 (04)
  • [45] Semantics of Multiword Expressions in Transformer-Based Models: A Survey
    Miletic, Filip
    Walde, Sabine Schulte Im
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 593 - 612
  • [46] Detecting microsatellite instability in colorectal cancer using Transformer-based colonoscopy image classification and retrieval
    Lo, Chung-Ming
    Jiang, Jeng-Kai
    Lin, Chun-Chi
    PLOS ONE, 2024, 19 (01):
  • [47] Are transformer-based models more robust than CNN-based models?
    Liu, Zhendong
    Qian, Shuwei
    Xia, Changhong
    Wang, Chongjun
    NEURAL NETWORKS, 2024, 172
  • [48] Strawberry disease identification with vision transformer-based models
    Nguyen, Hai Thanh
    Tran, Tri Dac
    Nguyen, Thanh Tuong
    Pham, Nhi Minh
    Nguyen Ly, Phuc Hoang
    Luong, Huong Hoang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (29) : 73101 - 73126
  • [49] Applications of transformer-based language models in bioinformatics: a survey
    Zhang, Shuang
    Fan, Rui
    Liu, Yuti
    Chen, Shuang
    Liu, Qiao
    Zeng, Wanwen
    NEURO-ONCOLOGY ADVANCES, 2023, 5 (01)
  • [50] TAG: Gradient Attack on Transformer-based Language Models
    Deng, Jieren
    Wang, Yijue
    Li, Ji
    Wang, Chenghong
    Shang, Chao
    Liu, Hang
    Rajasekaran, Sanguthevar
    Ding, Caiwen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3600 - 3610