VerilogEval: Evaluating Large Language Models for Verilog Code Generation

09/14/2023
by   Mingjie Liu, et al.
0

The increasing popularity of large language models (LLMs) has paved the way for their application in diverse domains. This paper proposes a benchmarking framework tailored specifically for evaluating LLM performance in the context of Verilog code generation for hardware design and verification. We present a comprehensive evaluation dataset consisting of 156 problems from the Verilog instructional website HDLBits. The evaluation set consists of a diverse set of Verilog code generation tasks, ranging from simple combinational circuits to complex finite state machines. The Verilog code completions can be automatically tested for functional correctness by comparing the transient simulation outputs of the generated design with a golden solution. We also demonstrate that the Verilog code generation capability of pretrained language models could be improved with supervised fine-tuning by bootstrapping with LLM generated synthetic problem-code pairs.

READ FULL TEXT

page 1

page 2

page 3

research
07/27/2023

PanGu-Coder2: Boosting Large Language Models for Code with Ranking Feedback

Large Language Models for Code (Code LLM) are flourishing. New and power...
research
12/13/2022

Benchmarking Large Language Models for Automated Verilog RTL Code Generation

Automating hardware design could obviate a significant amount of human e...
research
07/28/2023

VeriGen: A Large Language Model for Verilog Code Generation

In this study, we explore the capability of Large Language Models (LLMs)...
research
10/31/2022

A Simple, Yet Effective Approach to Finding Biases in Code Generation

Recently, scores of high-performing code generation systems have surface...
research
04/26/2023

A Deep Learning Framework for Verilog Autocompletion Towards Design and Verification Automation

Innovative Electronic Design Automation (EDA) solutions are important to...
research
02/10/2023

Large Language Models for Code: Security Hardening and Adversarial Testing

Large language models (LMs) are increasingly pretrained on massive codeb...
research
06/22/2023

DiversiGATE: A Comprehensive Framework for Reliable Large Language Models

In this paper, we introduce DiversiGATE, a unified framework that consol...

Please sign up or login with your details

Forgot password? Click here to reset