Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

VeriGen: A Large Language Model for Verilog Code Generation (2308.00708v1)

Published 28 Jul 2023 in cs.PL, cs.LG, and cs.SE

Abstract: In this study, we explore the capability of LLMs to automate hardware design by generating high-quality Verilog code, a common language for designing and modeling digital systems. We fine-tune pre-existing LLMs on Verilog datasets compiled from GitHub and Verilog textbooks. We evaluate the functional correctness of the generated Verilog code using a specially designed test suite, featuring a custom problem set and testing benches. Here, our fine-tuned open-source CodeGen-16B model outperforms the commercial state-of-the-art GPT-3.5-turbo model with a 1.1% overall increase. Upon testing with a more diverse and complex problem set, we find that the fine-tuned model shows competitive performance against state-of-the-art gpt-3.5-turbo, excelling in certain scenarios. Notably, it demonstrates a 41% improvement in generating syntactically correct Verilog code across various problem categories compared to its pre-trained counterpart, highlighting the potential of smaller, in-house LLMs in hardware design automation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shailja Thakur (12 papers)
  2. Baleegh Ahmad (9 papers)
  3. Hammond Pearce (35 papers)
  4. Benjamin Tan (42 papers)
  5. Brendan Dolan-Gavitt (24 papers)
  6. Ramesh Karri (92 papers)
  7. Siddharth Garg (99 papers)
Citations (84)