共 50 条
- [41] RoleLLM: Benchmarking, Eliciting, and Enhancing Role-Playing Abilities of Large Language Models FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 14743 - 14777
- [42] On the Evaluation of Large Language Models in Unit Test Generation Proceedings - 2024 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024, : 1607 - 1619
- [43] Updating knowledge in Large Language Models: an Empirical Evaluation IEEE CONFERENCE ON EVOLVING AND ADAPTIVE INTELLIGENT SYSTEMS 2024, IEEE EAIS 2024, 2024, : 289 - 296
- [45] A Comprehensive Evaluation of Quantization Strategies for Large Language Models FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 12186 - 12215
- [46] Can Large Language Models Be an Alternative to Human Evaluation? PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15607 - 15631
- [49] An Empirical Analysis on Large Language Models in Debate Evaluation PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 470 - 487
- [50] Benchmarking protein language models for protein crystallization SCIENTIFIC REPORTS, 2025, 15 (01):