VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation

被引:0
|
作者
Vijayaraghavan, Prashanth [1 ]
Shi, Luyao [1 ]
Ambrogio, Stefano [1 ]
Mackin, Charles [1 ]
Nitsure, Apoorva [1 ]
Beymer, David [1 ]
Degan, Ehsan [1 ]
机构
[1] IBM Res, San Jose, CA 95120 USA
关键词
LLMs; large language models; VHDL Code generation; VHDL Evaluation; hardware design automation; Hardware Description Languages; HDL; PEFT; ICL;
D O I
10.1109/LAD62341.2024.10691836
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the unprecedented advancements in Large Language Models (LLMs), their application domains have expanded to include code generation tasks across various programming languages. While significant progress has been made in enhancing LLMs for popular programming languages, there exists a notable gap in comprehensive evaluation frameworks tailored for Hardware Description Languages (HDLs), particularly VHDL. This paper addresses this gap by introducing a comprehensive evaluation framework designed specifically for assessing LLM performance in VHDL code generation task. We construct a dataset for evaluating LLMs on VHDL code generation task. This dataset is constructed by translating a collection of Verilog evaluation problems to VHDL and aggregating publicly available VHDL problems, resulting in a total of 202 problems. To assess the functional correctness of the generated VHDL code, we utilize a curated set of self-verifying testbenches specifically designed for those aggregated VHDL problem set. We conduct an initial evaluation of different LLMs and their variants, including zero-shot code generation, in-context learning (ICL), and Parameter-efficient fine-tuning (P EFT) methods. Our findings underscore the considerable challenges faced by existing LLMs in VHDL code generation, revealing significant scope for improvement. This study emphasizes the necessity of supervised fine-tuning code generation models specifically for VHDL, offering potential benefits to V HDL designers seeking efficient code generation solutions.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Systematic VHDL code generation using pipeline operations produced by high level synthesis
    Arató, P
    Kandár, T
    2003 IEEE INTERNATIONAL SYMPOSIUM ON INTELLIGENT SIGNAL PROCESSING, PROCEEDINGS: FROM CLASSICAL MEASUREMENT TO COMPUTING WITH PERCEPTIONS, 2003, : 191 - 196
  • [32] An evolutionary approach to automatic generation of VHDL code for low-power digital filters
    Erba, M
    Rossi, R
    Liberali, V
    Tettamanzi, GB
    GENETIC PROGRAMMING, PROCEEDINGS, 2001, 2038 : 36 - 50
  • [33] A Comparative Analysis of Large Language Models for Code Documentation Generation
    Dvivedi, Shubhang Shekhar
    Vijay, Vyshnav
    Pujari, Sai Leela Rahul
    Lodh, Shoumik
    Kumar, Dhruv
    PROCEEDINGS OF THE 1ST ACM INTERNATIONAL CONFERENCE ON AI-POWERED SOFTWARE, AIWARE 2024, 2024, : 65 - 73
  • [34] BioCoder: a benchmark for bioinformatics code generation with large language models
    Tang, Xiangru
    Qian, Bill
    Gao, Rick
    Chen, Jiakang
    Chen, Xinyun
    Gerstein, Mark B.
    BIOINFORMATICS, 2024, 40 : i266 - i276
  • [35] Knowledge-Aware Code Generation with Large Language Models
    Huang, Tao
    Sun, Zhihong
    Jin, Zhi
    Li, Ge
    Lyu, Chen
    PROCEEDINGS 2024 32ND IEEE/ACM INTERNATIONAL CONFERENCE ON PROGRAM COMPREHENSION, ICPC 2024, 2024, : 52 - 63
  • [36] Self-Planning Code Generation with Large Language Models
    Jiang, Xue
    Dong, Yihong
    Wang, Lecheng
    Fang, Zheng
    Shang, Qiwei
    Li, Ge
    Jin, Zhi
    Jiao, Wenpin
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (07)
  • [37] Evaluating Source Code Quality with Large Language Models: a comparative study
    da Silva Simões, Igor Regis
    Venson, Elaine
    arXiv,
  • [38] Evaluating Large Language Models for Enhanced Fuzzing: An Analysis Framework for LLM-Driven Seed Generation
    Black, Gavin
    Vaidyan, Varghese Mathew
    Comert, Gurcan
    IEEE ACCESS, 2024, 12 : 156065 - 156081
  • [39] CodeT5+: Open Code Large Language Models for Code Understanding and Generation
    Wang, Yue
    Le, Hung
    Gotmare, Akhilesh Deepak
    Bui, Nghi D. Q.
    Li, Junnan
    Hoi, Steven C. H.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 1069 - 1088
  • [40] GREEN-CODE: Optimizing Energy Efficiency in Large Language Models for Code Generation
    Ilager, Shashikant
    Briem, Lukas Florian
    Brandic, Ivona
    arXiv,