Large Language Model for Verilog Code Generation: Literature Review and the Road Ahead

📅 2025-10-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The absence of a systematic survey on large language models (LLMs) for Verilog RTL code generation hinders progress in hardware-AI co-design. Method: This paper conducts a cross-domain (software engineering/AI/EDA) analysis of 102 publications—70 peer-reviewed papers and 32 high-quality preprints—using bibliometric analysis, topic modeling, and cross-domain comparison, complemented by evaluation techniques including fine-tuning, prompt engineering, and functional correctness assessment. Contribution/Results: We introduce the first LLM-Verilog research map and a four-dimensional analytical framework addressing key research questions. Core bottlenecks are identified: architectural limitations of LLMs, scarcity of high-quality Verilog training data, and insufficient functional correctness guarantees. Furthermore, we propose a novel three-stage roadmap—“verifiable generation → hardware-aware alignment → EDA toolchain integration”—to guide future development, offering both theoretical foundations and practical pathways for LLM-driven RTL synthesis.

Technology Category

Application Category

📝 Abstract
Code generation has emerged as a critical research area at the intersection of Software Engineering (SE) and Artificial Intelligence (AI), attracting significant attention from both academia and industry. Within this broader landscape, Verilog, as a representative hardware description language (HDL), plays a fundamental role in digital circuit design and verification, making its automated generation particularly significant for Electronic Design Automation (EDA). Consequently, recent research has increasingly focused on applying Large Language Models (LLMs) to Verilog code generation, particularly at the Register Transfer Level (RTL), exploring how these AI-driven techniques can be effectively integrated into hardware design workflows. Despite substantial research efforts have explored LLM applications in this domain, a comprehensive survey synthesizing these developments remains absent from the literature. This review fill addresses this gap by providing a systematic literature review of LLM-based methods for Verilog code generation, examining their effectiveness, limitations, and potential for advancing automated hardware design. The review encompasses research work from conferences and journals in the fields of SE, AI, and EDA, encompassing 70 papers published on venues, along with 32 high-quality preprint papers, bringing the total to 102 papers. By answering four key research questions, we aim to (1) identify the LLMs used for Verilog generation, (2) examine the datasets and metrics employed in evaluation, (3) categorize the techniques proposed for Verilog generation, and (4) analyze LLM alignment approaches for Verilog generation. Based on our findings, we have identified a series of limitations of existing studies. Finally, we have outlined a roadmap highlighting potential opportunities for future research endeavors in LLM-assisted hardware design.
Problem

Research questions and friction points this paper is trying to address.

Surveying LLM applications for automated Verilog code generation
Analyzing effectiveness and limitations of AI-driven hardware design methods
Providing roadmap for future LLM-assisted electronic design automation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Systematic review of LLM-based Verilog generation methods
Analysis of techniques and alignment approaches for hardware design
Roadmap for future LLM-assisted automated hardware design
G
Guang Yang
Zhejiang University & Northwestern Polytechnical University, China
W
Wei Zheng
Northwestern Polytechnical University, China
X
Xiang Chen
Nantong University, China
D
Dong Liang
Northwestern Polytechnical University, China
P
Peng Hu
Northwestern Polytechnical University, China
Y
Yukui Yang
Northwestern Polytechnical University, China
S
Shaohang Peng
Northwestern Polytechnical University, China
Z
Zhenghan Li
Northwestern Polytechnical University, China
J
Jiahui Feng
Northwestern Polytechnical University, China
Xiao Wei
Xiao Wei
Duke University
roboticsrobot learningreinforcement learning
K
Kexin Sun
Northwestern Polytechnical University, China
D
Deyuan Ma
Northwestern Polytechnical University, China
H
Hao Cheng
Northwestern Polytechnical University, China
Yiheng Shen
Yiheng Shen
Computer Science, Duke University
Economics and Computation
X
Xing Hu
Zhejiang University, China
Terry Yue Zhuo
Terry Yue Zhuo
Researcher
Large Language ModelsCode GenerationAI4SECybersecurity
D
David Lo
Singapore Management University, Singapore