Transformer based contextualization of pre-trained word embeddings for irony detection in Twitter

被引:39
|
作者
Angel Gonzalez, Jose [1 ]
Hurtado, Lluis-F [1 ]
Pla, Ferran [1 ]
机构
[1] Univ Politecn Valencia, VRAIN Valencian Res Inst Artificial Intelligence, Cami Vera Sn, Valencia 46022, Spain
关键词
Irony detection; Twitter; Deep learning; Transformer encoders;
D O I
10.1016/j.ipm.2020.102262
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human communication using natural language, specially in social media, is influenced by the use of figurative language like irony. Recently, several workshops are intended to explore the task of irony detection in Twitter by using computational approaches. This paper describes a model for irony detection based on the contextualization of pre-trained Twitter word embeddings by means of the Transformer architecture. This approach is based on the same powerful architecture as BERT but, differently to it, our approach allows us to use in-domain embeddings. We performed an extensive evaluation on two corpora, one for the English language and another for the Spanish language. Our system was the first ranked system in the Spanish corpus and, to our knowledge, it has achieved the second-best result on the English corpus. These results support the correctness and adequacy of our proposal. We also studied and interpreted how the multi-head self-attention mechanisms are specialized on detecting irony by means of considering the polarity and relevance of individual words and even the relationships among words. This analysis is a first step towards understanding how the multi-head self-attention mechanisms of the Transformer architecture address the irony detection problem.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Pre-trained transformer-based language models for Sundanese
    Wilson Wongso
    Henry Lucky
    Derwin Suhartono
    [J]. Journal of Big Data, 9
  • [32] A survey of transformer-based multimodal pre-trained modals
    Han, Xue
    Wang, Yi-Tong
    Feng, Jun-Lan
    Deng, Chao
    Chen, Zhan-Heng
    Huang, Yu-An
    Su, Hui
    Hu, Lun
    Hu, Peng-Wei
    [J]. NEUROCOMPUTING, 2023, 515 : 89 - 106
  • [33] Pre-trained transformer-based language models for Sundanese
    Wongso, Wilson
    Lucky, Henry
    Suhartono, Derwin
    [J]. JOURNAL OF BIG DATA, 2022, 9 (01)
  • [34] Chemformer: a pre-trained transformer for computational chemistry
    Irwin, Ross
    Dimitriadis, Spyridon
    He, Jiazhen
    Bjerrum, Esben Jannik
    [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (01):
  • [35] PART: Pre-trained Authorship Representation Transformer
    Huertas-Tato, Javier
    Martin, Alejandro
    Camacho, David
    [J]. HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES, 2024, 14
  • [36] Integrally Pre-Trained Transformer Pyramid Networks
    Tian, Yunjie
    Xie, Lingxi
    Wang, Zhaozhi
    Wei, Longhui
    Zhang, Xiaopeng
    Jiao, Jianbin
    Wang, Yaowei
    Tian, Qi
    Ye, Qixiang
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18610 - 18620
  • [37] On the Role of Pre-trained Embeddings in Binary Code Analysis
    Maier, Alwin
    Weissberg, Felix
    Rieck, Konrad
    [J]. PROCEEDINGS OF THE 19TH ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, ACM ASIACCS 2024, 2024, : 795 - 810
  • [38] Pre-trained Embeddings for Entity Resolution: An Experimental Analysis
    Zeakis, Alexandros
    Papadakis, George
    Skoutas, Dimitrios
    Koubarakis, Manolis
    [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2023, 16 (09): : 2225 - 2238
  • [39] On the Sentence Embeddings from Pre-trained Language Models
    Li, Bohan
    Zhou, Hao
    He, Junxian
    Wang, Mingxuan
    Yang, Yiming
    Li, Lei
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9119 - 9130
  • [40] An integrated model based on deep learning classifiers and pre-trained transformer for phishing URL detection
    Do, Nguyet Quang
    Selamat, Ali
    Fujita, Hamido
    Krejcar, Ondrej
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 161 : 269 - 285