SiliconMind-V1: Multi-Agent Distillation and Debug-Reasoning Workflows for Verilog Code Generation

📅 2026-02-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing large language models in Verilog code generation, which often rely heavily on proprietary models or external verification tools, resulting in high costs, privacy concerns, and insufficient functional correctness. To overcome these challenges, the authors propose a unified multi-agent framework that leverages a testbench-driven verification mechanism to automatically generate inference-oriented training data. Combined with a test-time expansion strategy, this approach enables iterative generation, verification, and debugging of RTL designs entirely without external tools. Through local fine-tuning alone, the method significantly improves functional correctness. Evaluated on the VerilogEval-v2, RTLLM-v2, and CVDP benchmarks, it surpasses the current state-of-the-art model, QiMeng-CodeV-R1, while using fewer training resources.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have recently emerged as a promising approach for automating Verilog code generation; however, existing methods primarily emphasize syntactic correctness and often rely on commercial models or external verification tools, which introduces concerns regarding cost, data privacy, and limited guarantees of functional correctness. This work proposes a unified multi-agent framework for reasoning-oriented training data generation with integrated testbench-driven verification, enabling locally fine-tuned LLMs, SiliconMind-V1, to iteratively generate, test, and debug Register-Transfer Level (RTL) designs through test-time scaling. Experimental results on representative benchmarks (VerilogEval-v2, RTLLM-v2, and CVDP) demonstrate that the proposed approach outperforms the state-of-the-art QiMeng-CodeV-R1 in functional correctness while using fewer training resources.
Problem

Research questions and friction points this paper is trying to address.

Verilog code generation
functional correctness
large language models
data privacy
verification
Innovation

Methods, ideas, or system contributions that make the work stand out.

multi-agent distillation
debug-reasoning workflows
testbench-driven verification
test-time scaling
functional correctness
🔎 Similar Papers
No similar papers found.
M
Mu-Chi Chen
Academia Sinica, Taipei, Taiwan
Y
Yu-Hung Kao
National Taiwan University, Taipei, Taiwan
P
Po-Hsuan Huang
National Taiwan University, Taipei, Taiwan
S
Shao-Chun Ho
National Taiwan University, Taipei, Taiwan
H
Hsiang-Yu Tsou
National Taiwan University, Taipei, Taiwan
I
I-Ting Wu
National Taiwan University, Taipei, Taiwan
E
En-Ming Huang
National Taiwan University, Taipei, Taiwan
Y
Yu-Kai Hung
National Taiwan University, Taipei, Taiwan
W
Wei-Po Hsin
National Taiwan University, Taipei, Taiwan
Cheng Liang
Cheng Liang
Shanghai AI Lab
VLM
Chia-Heng Tu
Chia-Heng Tu
National Cheng Kung University
Heterogeneous parallel computingEmbedded systems design and optimizationCompiler design
Shih-Hao Hung
Shih-Hao Hung
National Taiwan University
Computer ArchitectureParallel ComputingPerformanceVirtualizationGPU
H. T. Kung
H. T. Kung
Professor, Harvard University
Machine learning acceleratorshigh-performance computingcomputer & wireless networkscomplexitydatabase systems