Large Processor Chip Model

📅 2025-06-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Conventional architecture design relies heavily on manual expertise, suffers from siloed hardware-software optimization, and incurs prohibitively high costs for design-space exploration. Method: This paper proposes LPCM, an LLM-driven three-tier collaborative framework that establishes the first “human-agent-model” co-design paradigm, deeply embedding large language models into a closed-loop hardware-software co-design workflow to overcome limitations of single-stage and fragmented optimization. It innovatively integrates 3D Gaussian splatting–based workload modeling with system-level co-design methodology. Contribution/Results: At Level 1 validation, LPCM achieves full automation of the end-to-end architecture design pipeline. Experiments demonstrate substantial reduction in design cycle time and human effort, establishing a scalable, reusable technical pathway toward fully autonomous, full-stack chip design.

Technology Category

Application Category

📝 Abstract
Computer System Architecture serves as a crucial bridge between software applications and the underlying hardware, encompassing components like compilers, CPUs, coprocessors, and RTL designs. Its development, from early mainframes to modern domain-specific architectures, has been driven by rising computational demands and advancements in semiconductor technology. However, traditional paradigms in computer system architecture design are confronting significant challenges, including a reliance on manual expertise, fragmented optimization across software and hardware layers, and high costs associated with exploring expansive design spaces. While automated methods leveraging optimization algorithms and machine learning have improved efficiency, they remain constrained by a single-stage focus, limited data availability, and a lack of comprehensive human domain knowledge. The emergence of large language models offers transformative opportunities for the design of computer system architecture. By leveraging the capabilities of LLMs in areas such as code generation, data analysis, and performance modeling, the traditional manual design process can be transitioned to a machine-based automated design approach. To harness this potential, we present the Large Processor Chip Model (LPCM), an LLM-driven framework aimed at achieving end-to-end automated computer architecture design. The LPCM is structured into three levels: Human-Centric; Agent-Orchestrated; and Model-Governed. This paper utilizes 3D Gaussian Splatting as a representative workload and employs the concept of software-hardware collaborative design to examine the implementation of the LPCM at Level 1, demonstrating the effectiveness of the proposed approach. Furthermore, this paper provides an in-depth discussion on the pathway to implementing Level 2 and Level 3 of the LPCM, along with an analysis of the existing challenges.
Problem

Research questions and friction points this paper is trying to address.

Addressing manual expertise reliance in architecture design
Overcoming fragmented software-hardware optimization challenges
Reducing high costs in expansive design space exploration
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM-driven automated computer architecture design
Three-level framework: Human, Agent, Model
Software-hardware collaborative design approach
🔎 Similar Papers
No similar papers found.
K
Kaiyan Chang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
M
Mingzhi Chen
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Yunji Chen
Yunji Chen
Institute of Computing Technology, Chinese Academy of Sciences
processor architecturemicroarchitecturemachine learning
Zhirong Chen
Zhirong Chen
Master, Institute of Computing Technology, Chinese Academy of Sciences
Computer ArchitectureMachine Learning
Dongrui Fan
Dongrui Fan
Institute of Computing Technology, Chinese Academy of Sciences
Computer ArchitectureProcessor DesignMany-core Design
J
Junfeng Gong
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
N
Nan Guo
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Y
Yinhe Han
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Q
Qinfen Hao
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
S
Shuo Hou
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
X
Xuan Huang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
P
Pengwei Jin
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
C
Changxin Ke
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
C
Cangyuan Li
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
G
Guangli Li
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Huawei Li
Huawei Li
Institute of Computing Technology, Chinese Academy of Sciences
computer engineering
Kuan Li
Kuan Li
Hong Kong University of Science and Technology (HKUST)
LLM agentmachine learning on graphsadversarial robustness
Naipeng Li
Naipeng Li
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Shengwen Liang
Shengwen Liang
Institute of computing technology, Chinese Academy of Sciences
AcceleratorCognitive SSDSystem
C
Cheng Liu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
H
Hongwei Liu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
J
Jiahua Liu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
J
Junliang Lv
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Jianan Mu
Jianan Mu
Institute of Computing Technology, State Key Laboratory of Processors (SKLP), CAS
Design AutomationAccelaretorPrivacy Preserving Computing
J
Jin Qin
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
B
Bin Sun
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
C
Chenxi Wang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
D
Duo Wang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Mingjun Wang
Mingjun Wang
Xidian University
Information SecurityWireless SecurityMobile Social Networking
Y
Ying Wang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
C
Chen-Chen Wu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
P
Peiyang Wu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
T
Teng Wu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
X
Xiao Xiao
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
M
Mengyao Xie
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
C
Chenwei Xiong
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
R
Ruiyuan Xu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
M
Mingyu Yan
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
X
Xiaochun Ye
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
K
Kuai Yu
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
R
Rui Zhang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
S
Shuoming Zhang
Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China
Jiacheng Zhao
Jiacheng Zhao
Institute of Computing Technology, Chinese Academy of Scienses
Parallel ComputingParallel CompilingComputer ArchitectureProgramming ModelDatacenter