CodeT5+: Open Code Large Language Models for Code Understanding and Generation

被引:0
|
作者
Wang, Yue [1 ]
Le, Hung [1 ]
Gotmare, Akhilesh Deepak [1 ]
Bui, Nghi D. Q. [1 ]
Li, Junnan [1 ]
Hoi, Steven C. H. [1 ]
机构
[1] Salesforce AI Res, San Francisco, CA 94105 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) pretrained on vast source code have achieved prominent progress in code intelligence. However, existing code LLMs have two main limitations. First, they often adopt a specific architecture (encoder-only or decoder-only) or rely on a unified encoder-decoder network for different downstream tasks, lacking the flexibility to operate in the optimal architecture for a specific task. Secondly, they often employ a limited set of pretraining objectives which might not be relevant to some tasks and hence result in substantial performance degrade. To address these limitations, we propose "CodeT5+", a family of encoder-decoder LLMs for code in which component modules can be flexibly combined to suit a wide range of code tasks. Such flexibility is enabled by our proposed mixture of pretraining objectives, which cover span denoising, contrastive learning, text-code matching, and causal LM pretraining tasks, on both unimodal and bimodal multilingual code corpora. Furthermore, we propose to initialize CodeT5+ with frozen off-the-shelf LLMs without training from scratch to efficiently scale up our models, and explore instruction-tuning to align with natural language instructions. We extensively evaluate CodeT5+ on over 20 code-related benchmarks in different settings, including zero-shot, finetuning, and instruction-tuning. We observe state-of-the-art (SoTA) performance on various code-related tasks, and our instruction-tuned CodeT5+ 16B achieves new SoTA results of 35.0% pass@1 and 54.5% pass@10 on the HumanEval code generation task against other open code LLMs, even surpassing the OpenAI code-cushman-001 model.
引用
收藏
页码:1069 / 1088
页数:20
相关论文
共 50 条
  • [21] Automatic Generation of Programming Exercises and Code Explanations Using Large Language Models
    Sarsa, Sami
    Denny, Paul
    Hellas, Arto
    Leinonen, Juho
    PROCEEDINGS OF THE 2022 ACM CONFERENCE ON INTERNATIONAL COMPUTING EDUCATION RESEARCH, ICER 2022, VOL. 1, 2023, : 27 - 43
  • [22] Hot or Cold? Adaptive Temperature Sampling for Code Generation with Large Language Models
    Zhu, Yuqi
    Li, Jia
    Li, Ge
    Zhao, YunFei
    Li, Jia
    Jin, Zhi
    Mei, Hong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 1, 2024, : 437 - 445
  • [23] VeriGen: A Large Language Model for Verilog Code Generation
    Thakur, Shailja
    Ahmad, Baleegh
    Pearce, Hammond
    Tan, Benjamin
    Dolan-Gavitt, Brendan
    Karri, Ramesh
    Garg, Siddharth
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2024, 29 (03)
  • [24] Structural language models for any-code generation
    Alon, Uri
    Sadaka, Roy
    Levy, Omer
    Yahav, Eran
    arXiv, 2019,
  • [25] JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models
    Cao, Jialun
    Chen, Zhiyong
    Wu, Jiarong
    Cheung, Shing-Chi
    Xu, Chang
    Proceedings - 2024 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024, : 870 - 882
  • [26] Enhancing Large Language Models-Based Code Generation by Leveraging Genetic Improvement
    Pinna, Giovanni
    Ravalico, Damiano
    Rovito, Luigi
    Manzoni, Luca
    De Lorenzo, Andrea
    GENETIC PROGRAMMING, EUROGP 2024, 2024, 14631 : 108 - 124
  • [27] Can ChatGPT Support Developers? An Empirical Evaluation of Large Language Models for Code Generation
    Jin, Kailun
    Wang, Chung-Yu
    Hung Viet Pham
    Hemmati, Hadi
    2024 IEEE/ACM 21ST INTERNATIONAL CONFERENCE ON MINING SOFTWARE REPOSITORIES, MSR, 2024, : 167 - 171
  • [28] VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation
    Vijayaraghavan, Prashanth
    Shi, Luyao
    Ambrogio, Stefano
    Mackin, Charles
    Nitsure, Apoorva
    Beymer, David
    Degan, Ehsan
    2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024, 2024,
  • [29] Balancing Security and Correctness in Code Generation: An Empirical Study on Commercial Large Language Models
    Black, Gavin S.
    Rimal, Bhaskar P.
    Vaidyan, Varghese Mathew
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (01): : 419 - 430
  • [30] A Comprehensive Framework for Evaluating API-oriented Code Generation in Large Language Models
    Wu, Yixi
    He, Pengfei
    Wang, Zehao
    Wang, Shaowei
    Tian, Yuan
    Chen, Tse-Hsun
    arXiv,