Incorporating Pre-trained Transformer Models into TextCNN for Sentiment Analysis on Software Engineering Texts

被引:2
|
作者
Sun, Kexin [1 ]
Shi, XiaoBo [2 ]
Gao, Hui [1 ]
Kuang, Hongyu [1 ]
Ma, Xiaoxing [1 ]
Rong, Guoping [1 ]
Shao, Dong [1 ]
Zhao, Zheng [3 ]
Zhang, He [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Dalian Maritime Univ, Coll Informat Sci & Technol, Dalian, Peoples R China
[3] Dalian Maritime Univ, Coll Artificial Intelligence, Dalian, Peoples R China
基金
中国国家自然科学基金;
关键词
Sentiment Analysis; Pre-trained Models; Software Mining; Nature Language Processing;
D O I
10.1145/3545258.3545273
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Software information sites (e.g., Jira, Stack Overflow) are now widely used in software development. These online platforms for collaborative development preserve a large amount of Software Engineering (SE) texts. These texts enable researchers to detect developers' attitudes toward their daily development by analyzing the sentiments expressed in the texts. Unfortunately, recent works reported that neither off-the-shelf tools nor SE-specified tools for sentiment analysis on SE texts can provide satisfying and reliable results. In this paper, we propose to incorporate pre-trained transformer models into the sentence-classification oriented deep learning framework named TextCNN to better capture the unique expression of sentiments in SE texts. Specifically, we introduce an optimized BERT model named RoBERTa as the word embedding layer of TextCNN, along with additional residual connections between RoBERTa and TextCNN for better cooperation in our training framework. An empirical evaluation based on four datasets from different software information sites shows that our training framework can achieve overall better accuracy and generalizability than the four baselines.
引用
收藏
页码:127 / 136
页数:10
相关论文
共 50 条
  • [1] Sentiment Analysis for Software Engineering: How Far Can Pre-trained Transformer Models Go?
    Zhang, Ting
    Xu, Bowen
    Thung, Ferdian
    Haryono, Stefanus Agus
    Lo, David
    Jiang, Lingxiao
    2020 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME 2020), 2020, : 70 - 80
  • [2] Emotion Classification In Software Engineering Texts: A Comparative Analysis of Pre-trained Transformers Language Models
    Imran, Mia Mohammad
    PROCEEDINGS 2024 ACM/IEEE INTERNATIONAL WORKSHOP ON NL-BASED SOFTWARE ENGINEERING, NLBSE 2024, 2024, : 73 - 80
  • [3] Incorporating emoji sentiment information into a pre-trained language model for Chinese and English sentiment analysis
    Huang, Jiaming
    Li, Xianyong
    Li, Qizhi
    Du, Yajun
    Fan, Yongquan
    Chen, Xiaoliang
    Huang, Dong
    Wang, Shumin
    Li, Xianyong
    INTELLIGENT DATA ANALYSIS, 2024, 28 (06) : 1601 - 1625
  • [4] Deep Compression of Pre-trained Transformer Models
    Wang, Naigang
    Liu, Chi-Chun
    Venkataramani, Swagath
    Sen, Sanchari
    Chen, Chia-Yu
    El Maghraoui, Kaoutar
    Srinivasan, Vijayalakshmi
    Chang, Leland
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [5] Enhancing Turkish Sentiment Analysis Using Pre-Trained Language Models
    Koksal, Omer
    29TH IEEE CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS (SIU 2021), 2021,
  • [6] Aspect Based Sentiment Analysis using French Pre-Trained Models
    Essebbar, Abderrahman
    Kane, Bamba
    Guinaudeau, Ophelie
    Chiesa, Valeria
    Quenel, Ilhem
    Chau, Stephane
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 1, 2021, : 519 - 525
  • [7] Detecting Syntactic Change with Pre-trained Transformer Models
    Hou, Liwen
    Smith, David A.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3564 - 3574
  • [8] On the effect of dropping layers of pre-trained transformer models
    Sajjad, Hassan
    Dalvi, Fahim
    Durrani, Nadir
    Nakov, Preslav
    COMPUTER SPEECH AND LANGUAGE, 2022, 77
  • [9] Incorporating Dynamic Semantics into Pre-Trained Language Model for Aspect-based Sentiment Analysis
    Zhang, Kai
    Zhang, Kun
    Zhang, Mengdi
    Zhao, Hongke
    Liu, Qi
    Wu, Wei
    Chen, Enhong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3599 - 3610
  • [10] Devulgarization of Polish Texts Using Pre-trained Language Models
    Klamra, Cezary
    Wojdyga, Grzegorz
    Zurowski, Sebastian
    Rosalska, Paulina
    Kozlowska, Matylda
    Ogrodniczuk, Maciej
    COMPUTATIONAL SCIENCE, ICCS 2022, PT II, 2022, : 49 - 55