共 50 条
- [2] JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models Proceedings - 2024 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024, : 870 - 882
- [4] CodeBERTScore: Evaluating Code Generation with Pretrained Models of Code 2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 13921 - 13937
- [7] Invited Paper: VerilogEval: Evaluating Large Language Models for Verilog Code Generation 2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
- [9] Evaluating Large Language Models for G-Code Debugging, Manipulation, and Comprehension 2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024, 2024,
- [10] Evaluating and Optimizing the Effectiveness of Neural Machine Translation in Supporting Code Retrieval Models: A Study on the CAT Benchmark International Conference on Information and Knowledge Management, Proceedings, 2023, : 2055 - 2064