Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BetterV: Controlled Verilog Generation with Discriminative Guidance (2402.03375v3)

Published 3 Feb 2024 in cs.AI and cs.PL

Abstract: Due to the growing complexity of modern Integrated Circuits (ICs), there is a need for automated circuit design methods. Recent years have seen rising research in hardware design language generation to facilitate the design process. In this work, we propose a Verilog generation framework, BetterV, which fine-tunes the LLMs on processed domain-specific datasets and incorporates generative discriminators for guidance on particular design demands. The Verilog modules are collected, filtered and processed from internet to form a clean and abundant dataset. Instruct-tuning methods are specially designed to fine-tune the LLMs to understand the knowledge about Verilog. Furthermore, data are augmented to enrich the training set and also used to train a generative discriminator on particular downstream task, which leads a guidance for the LLMs to optimize the Verilog implementation. BetterV has the ability to generate syntactically and functionally correct Verilog, which can outperform GPT-4 on the VerilogEval benchmark. With the help of task-specific generative discriminator, BetterV can achieve remarkable improvement on various electronic design automation (EDA) downstream tasks, including the netlist node reduction for synthesis and verification runtime reduction with Boolean Satisfiability (SAT) solving.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Evaluating large language models trained on code. arXiv preprint arXiv:2107.03374, 2021.
  2. Plug and Play Language Models: A Simple Approach to Controlled Text Generation. In International Conference on Learning Representations (ICLR), 2020.
  3. A Deep Learning Framework for Verilog Autocompletion Towards Design and Verification Automation. arXiv preprint arXiv:2304.13840, 2023.
  4. ChatEDA: A large language model powered autonomous agent for EDA. In ACM/IEEE Workshop on Machine Learning CAD (MLCAD), pp.  1–6. IEEE, 2023.
  5. Learning to write with cooperative discriminators. arXiv preprint arXiv:1805.06087, 2018.
  6. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
  7. Ctrl: A conditional transformer language model for controllable generation. arXiv preprint arXiv:1909.05858, 2019.
  8. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  9. Gedi: Generative discriminator guided sequence generation. arXiv preprint arXiv:2009.06367, 2020.
  10. DExperts: Decoding-time controlled text generation with experts and anti-experts. arXiv preprint arXiv:2105.03023, 2021.
  11. ChipNeMo: Domain-Adapted LLMs for Chip Design. arXiv preprint arXiv:2311.00176, 2023a.
  12. VerilogEval: Evaluating Large Language Models for Verilog Code Generation. In IEEE/ACM International Conference on Computer-Aided Design (ICCAD), pp.  1–8. IEEE, 2023b.
  13. RTLCoder: Outperforming GPT-3.5 in Design RTL Generation with Our Open-Source Dataset and Lightweight Solution. arXiv preprint arXiv:2312.08617, 2023c.
  14. Sgdr: Stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983, 2016.
  15. RTLLM: An open-source benchmark for design rtl generation with large language model. arXiv preprint arXiv:2308.05345, 2023.
  16. Hardware verification using software analyzers. In IEEE Computer Society Annual Symposium on VLSI, pp.  7–12. IEEE, 2015. ISBN 978-1-4799-8719-1.
  17. v2c – a Verilog to C translator tool. In Tools and Algorithms for the Construction and Analysis of Systems (TACAS), volume 9636 of LNCS, pp.  580–586. Springer, 2016. ISBN 978-3-662-49673-2.
  18. Codegen: An open large language model for code with multi-turn program synthesis. arXiv preprint arXiv:2203.13474, 2022.
  19. Codegen2: Lessons for training llms on programming and natural languages. arXiv preprint arXiv:2305.02309, 2023.
  20. Llm4eda: Emerging progress in large language models for electronic design automation. arXiv:2401.12224, 2023.
  21. Zero: Memory optimizations toward training trillion parameter models. In ACM/IEEE Supercomputing Conference (SC), pp.  1–16. IEEE, 2020.
  22. Code llama: Open foundation models for code. arXiv preprint arXiv:2308.12950, 2023.
  23. Discriminative adversarial search for abstractive summarization. In International Conference on Machine Learning (ICML), pp.  8555–8564. PMLR, 2020.
  24. Benchmarking Large Language Models for Automated Verilog RTL Code Generation. In IEEE/ACM Proceedings Design, Automation and Test in Eurpoe (DATE), pp.  1–6. IEEE, 2023.
  25. RTLFixer: Automatically Fixing RTL Syntax Errors with Large Language Models. arXiv preprint arXiv:2311.16543, 2023.
  26. Wolf, C. Yosys open synthesis suite. https://yosyshq.net/yosys/.
  27. DeepSpeed-Chat: Easy, Fast and Affordable RLHF Training of ChatGPT-like Models at All Scales. arXiv preprint arXiv:2308.01320, 2023.
  28. A survey of controllable text generation using transformer-based pre-trained language models. ACM Computing Surveys, 56(3):1–37, 2023.
  29. TinyLlama: An Open-Source Small Language Model, 2024.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Zehua Pei (7 papers)
  2. Hui-Ling Zhen (33 papers)
  3. Mingxuan Yuan (81 papers)
  4. Yu Huang (176 papers)
  5. Bei Yu (113 papers)
Citations (26)
X Twitter Logo Streamline Icon: https://streamlinehq.com