Can ChatGPT Support Developers? An Empirical Evaluation of Large Language Models for Code Generation

被引:3
|
作者
Jin, Kailun [1 ]
Wang, Chung-Yu [1 ]
Hung Viet Pham [1 ]
Hemmati, Hadi [1 ]
机构
[1] York Univ, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
10.1145/3643991.3645074
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) have demonstrated notable proficiency in code generation, with numerous prior studies showing their promising capabilities in various development scenarios. However, these studies mainly provide evaluations in research settings, which leaves a significant gap in understanding how effectively LLMs can support developers in real-world. To address this, we conducted an empirical analysis of conversations in DevGPT, a dataset collected from developers' conversations with ChatGPT (captured with the Share Link feature on platforms such as GitHub). Our empirical findings indicate that the current practice of using LLM-generated code is typically limited to either demonstrating high-level concepts or providing examples in documentation, rather than to be used as production-ready code. These findings indicate that there is much future work needed to improve LLMs in code generation before they can be integral parts of modern software development.
引用
收藏
页码:167 / 171
页数:5
相关论文
共 50 条
  • [21] Can Large Language Models Provide Feedback to Students? A Case Study on ChatGPT
    Dai, Wei
    Lin, Jionghao
    Jin, Hua
    Li, Tongguang
    Tsai, Yi-Shan
    Gasevic, Dragan
    Chen, Guanliang
    2023 IEEE INTERNATIONAL CONFERENCE ON ADVANCED LEARNING TECHNOLOGIES, ICALT, 2023, : 323 - 325
  • [22] Updating knowledge in Large Language Models: an Empirical Evaluation
    Marinelli, Alberto Roberto
    Carta, Antonio
    Passaro, Lucia C.
    IEEE CONFERENCE ON EVOLVING AND ADAPTIVE INTELLIGENT SYSTEMS 2024, IEEE EAIS 2024, 2024, : 289 - 296
  • [23] An Empirical Analysis on Large Language Models in Debate Evaluation
    Liu, Xinyi
    Liu, Pinxin
    He, Hangfeng
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 470 - 487
  • [24] An Empirical Study of the Non-Determinism of ChatGPT in Code Generation
    Ouyang, Shuyin
    Zhang, Jie m.
    Harman, Mark
    Wang, Meng
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2025, 34 (02)
  • [25] CodeT5+: Open Code Large Language Models for Code Understanding and Generation
    Wang, Yue
    Le, Hung
    Gotmare, Akhilesh Deepak
    Bui, Nghi D. Q.
    Li, Junnan
    Hoi, Steven C. H.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 1069 - 1088
  • [26] GREEN-CODE: Optimizing Energy Efficiency in Large Language Models for Code Generation
    Ilager, Shashikant
    Briem, Lukas Florian
    Brandic, Ivona
    arXiv,
  • [27] Evaluation of Large Language Models on Code Obfuscation (Student Abstract)
    Swindle, Adrian
    McNealy, Derrick
    Krishnan, Giri
    Ramyaa, Ramyaa
    THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23664 - 23666
  • [28] ARCHCODE: Incorporating Software Requirements in Code Generation with Large Language Models
    Han, Hojae
    Kim, Jaejin
    Yoo, Jaeseok
    Lee, Youngwon
    Hwang, Seung-won
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 13520 - 13552
  • [29] Automatic Unit Test Code Generation Using Large Language Models
    Ocal, Akdeniz Kutay
    Keskinoz, Mehmet
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [30] Harnessing the Power of Large Language Models for Automated Code Generation and Verification
    Antero, Unai
    Blanco, Francisco
    Onativia, Jon
    Salle, Damien
    Sierra, Basilio
    ROBOTICS, 2024, 13 (09)