Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Multi-Expert Large Language Model Architecture for Verilog Code Generation (2404.08029v1)

Published 11 Apr 2024 in cs.LG, cs.AI, cs.PL, and cs.SE

Abstract: Recently, there has been a surging interest in using LLMs for Verilog code generation. However, the existing approaches are limited in terms of the quality of the generated Verilog code. To address such limitations, this paper introduces an innovative multi-expert LLM architecture for Verilog code generation (MEV-LLM). Our architecture uniquely integrates multiple LLMs, each specifically fine-tuned with a dataset that is categorized with respect to a distinct level of design complexity. It allows more targeted learning, directly addressing the nuances of generating Verilog code for each category. Empirical evidence from experiments highlights notable improvements in terms of the percentage of generated Verilog outputs that are syntactically and functionally correct. These findings underscore the efficacy of our approach, promising a forward leap in the field of automated hardware design through machine learning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Dessouky et al. HardFails: Insights into Software-Exploitable hardware bugs. In 28th USENIX Security Symposium (USENIX Security 19), pages 213–230, Santa Clara, CA, August 2019. USENIX Association.
  2. Vaswani et al. Attention is all you need. 2017.
  3. Radford et al. Improving language understanding by generative pre-training. 2018.
  4. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  5. Cohen et al. Lamda: Language models for dialog applications. In arXiv. 2022.
  6. Thakur et al. Benchmarking large language models for automated verilog rtl code generation. 2023 Design, Automation & Test in Europe Conference & Exhibition (DATE), pages 1–6, 2022.
  7. Invited paper: Verilogeval: Evaluating large language models for verilog code generation. In 2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD), pages 1–8, 2023.
  8. Chang et al. Chipgpt: How far are we from natural language hardware design, 2023.
  9. Rtllm: An open-source benchmark for design rtl generation with large language model, 2023.
  10. Chip-chat: Challenges and opportunities in conversational hardware design. In 2023 ACM/IEEE 5th Workshop on Machine Learning for CAD (MLCAD). IEEE, September 2023.
  11. Fu et al. Gpt4aigchip: Towards next-generation ai accelerator design automation via large language models, 2023.
  12. Liu et al. Chipnemo: Domain-adapted llms for chip design, 2023.
  13. Rtlfixer: Automatically fixing rtl syntax errors with large language models, 2024.
  14. Thakur et al. Autochip: Automating hdl generation using llm feedback, 2023.
  15. Liu et al. Rtlcoder: Outperforming gpt-3.5 in design rtl generation with our open-source dataset and lightweight solution, 2024.
  16. Chang et al. Improving large language model hardware generating quality through post-LLM search. In Machine Learning for Systems 2023, 2023.
  17. System-on-chip message flow mining with masked-language models. In 2023 IEEE 66th International Midwest Symposium on Circuits and Systems (MWSCAS), pages 496–500, 2023.
  18. When llm-based code generation meets the software development process, 2024.
  19. Nijkamp et al. Codegen: An open large language model for code with multi-turn program synthesis. In The Eleventh International Conference on Learning Representations, 2023.
  20. Mesnard et al. Gemma: Open models based on gemini research and technology, 2024.
  21. A survey on data selection for llm instruction tuning, 2024.
  22. Curriculum labeling: Revisiting pseudo-labeling for semi-supervised learning. Proceedings of the AAAI Conference on Artificial Intelligence, 35(8):6912–6920, May 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Bardia Nadimi (8 papers)
  2. Hao Zheng (200 papers)
Citations (3)