GSIM: Accelerating RTL Simulation for Large-Scale Designs

📅 2025-08-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
RTL simulation is critical in hardware design, yet software-based simulation suffers from prohibitively low throughput for complex designs. This paper proposes a multi-level co-optimization methodology that systematically reduces four major computational overhead sources—across supernode, node, and bit-level abstractions—to build GSIM, an efficient open-source RTL simulator. Our approach innovatively integrates supernode-level scheduling, fine-grained event management, and bit-level computation optimization, while重构 the simulation engine via combined static analysis and runtime adaptive strategies. Experimental evaluation demonstrates that GSIM successfully simulates Linux boot on the XiangShan processor, achieving 7.34× speedup over Verilator; on the Rocket core, it attains 19.94× acceleration in CoreMark benchmarking. These results represent an order-of-magnitude performance improvement for large-scale RISC-V designs, establishing GSIM as a scalable, high-throughput RTL simulation framework.

Technology Category

Application Category

📝 Abstract
Register Transfer Level (RTL) simulation is widely used in design space exploration, verification, debugging, and preliminary performance evaluation for hardware design. Among various RTL simulation approaches, software simulation is the most commonly used due to its flexibility, low cost, and ease of debugging. However, the slow simulation of complex designs has become the bottleneck in design flow. In this work, we explore the sources of computation overhead of RTL simulation and conclude them into four factors. To optimize these factors, we propose several techniques at the supernode level, node level, and bit level. Finally, we implement these techniques in a novel RTL simulator GSIM. GSIM succeeds in simulating XiangShan, the state-of-the-art open-source RISC-V processor. Besides, compared to Verilator, GSIM can achieve speedup of 7.34x for booting Linux on XiangShan, and 19.94x for running CoreMark on Rocket.
Problem

Research questions and friction points this paper is trying to address.

Accelerating slow RTL simulation for complex hardware designs
Reducing computation overhead in RTL simulation techniques
Improving simulation speed for large-scale RISC-V processors
Innovation

Methods, ideas, or system contributions that make the work stand out.

Optimizes RTL simulation at supernode level
Enhances efficiency at node and bit levels
Implements GSIM for significant speedup
🔎 Similar Papers
No similar papers found.
L
Lu Chen
State Key Lab of Processors, Institute of Computing Technology, Chinese Academy of Sciences, China; University of Chinese Academy of Sciences, China
D
Dingyi Zhao
Beijing Institute of Open Source Chip, China
Zihao Yu
Zihao Yu
University of Science and Technology of China
N
Ninghui Sun
State Key Lab of Processors, Institute of Computing Technology, Chinese Academy of Sciences, China; University of Chinese Academy of Sciences, China
Yungang Bao
Yungang Bao
Institute of Computing Technology (ICT), CAS
Computer ArchitectureComputer System