An Empirical Comparison of Pre-Trained Models of Source Code

被引:18
|
作者
Niu, Changan [1 ]
Li, Chuanyi [1 ]
Ng, Vincent [2 ]
Chen, Dongxiao [1 ]
Ge, Jidong [1 ]
Luo, Bin [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Univ Texas Dallas, Human Language Technol Res Inst, Richardson, TX 75080 USA
基金
中国国家自然科学基金;
关键词
Pre-training of Source Code; AI for SE;
D O I
10.1109/ICSE48619.2023.00180
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
While a large number of pre-trained models of source code have been successfully developed and applied to a variety of software engineering (SE) tasks in recent years, our understanding of these pre-trained models is arguably fairly limited. With the goal of advancing our understanding of these models, we perform the first systematic empirical comparison of 19 recently-developed pre-trained models of source code on 13 SE tasks. To gain additional insights into these models, we adopt a recently-developed 4-dimensional categorization of pre-trained models, and subsequently investigate whether there are correlations between different categories of pre-trained models and their performances on different SE tasks.
引用
收藏
页码:2136 / 2148
页数:13
相关论文
共 50 条
  • [31] Pre-trained Models for Sonar Images
    Valdenegro-Toro, Matias
    Preciado-Grijalva, Alan
    Wehbe, Bilal
    OCEANS 2021: SAN DIEGO - PORTO, 2021,
  • [32] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    ENGINEERING, 2023, 25 : 51 - 65
  • [33] An Empirical Study on Robustness to Spurious Correlations using Pre-trained Language Models
    Tu, Lifu
    Lalwani, Garima
    Gella, Spandana
    He, He
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 : 621 - 633
  • [34] SCC-GPT: Source Code Classification Based on Generative Pre-Trained Transformers
    Alahmadi, Mohammad D.
    Alshangiti, Moayad
    Alsubhi, Jumana
    MATHEMATICS, 2024, 12 (13)
  • [35] CodeBERT-Attack: Adversarial attack against source code deep learning models via pre-trained model
    Zhang, Huangzhao
    Lu, Shuai
    Li, Zhuo
    Jin, Zhi
    Ma, Lei
    Liu, Yang
    Li, Ge
    JOURNAL OF SOFTWARE-EVOLUTION AND PROCESS, 2024, 36 (03)
  • [36] Improving Fine-tuning Pre-trained Models on Small Source Code Datasets via Variational Information Bottleneck
    Liu, Jiaxing
    Sha, Chaofeng
    Peng, Xin
    2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING, SANER, 2023, : 331 - 342
  • [37] An Open Medical Platform to Share Source Code and Various Pre-Trained Weights for Models to Use in Deep Learning Research
    Kim, Sungchul
    Cho, Sungman
    Cho, Kyungjin
    Seo, Jiyeon
    Nam, Yujin
    Park, Jooyoung
    Kim, Kyuri
    Kim, Daeun
    Hwang, Jeongeun
    Yun, Jihye
    Jang, Miso
    Lee, Hyunna
    Kim, Namkug
    KOREAN JOURNAL OF RADIOLOGY, 2021, 22 (12) : 2073 - 2081
  • [38] CODEFUSION: A Pre-trained Diffusion Model for Code Generation
    Singh, Mukul
    Cambronero, Jose
    Gulwani, Sumit
    Le, Vu
    Negreanu, Carina
    Verbruggen, Gust
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 11697 - 11708
  • [39] On the Role of Pre-trained Embeddings in Binary Code Analysis
    Maier, Alwin
    Weissberg, Felix
    Rieck, Konrad
    PROCEEDINGS OF THE 19TH ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, ACM ASIACCS 2024, 2024, : 795 - 810
  • [40] Bridge and Hint: Extending Pre-trained Language Models for Long-Range Code
    Chen, Yujia
    Gao, Cuiyun
    Yang, Zezhou
    Zhang, Hongyu
    Liao, Qing
    PROCEEDINGS OF THE 33RD ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2024, 2024, : 274 - 286