GLM-5: from Vibe Coding to Agentic Engineering

📅 2026-02-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes a paradigm shift in large language models—from intuition-driven “atmospheric encoding” to an agent-based framework endowed with autonomous reasoning and engineering capabilities—to effectively support complex, long-horizon software development tasks. To this end, we introduce GLM-5, a next-generation foundation model that incorporates an asynchronous reinforcement learning architecture and a novel agent-centric RL algorithm to decouple generation from training. Coupled with Dynamic Sparse Attention (DSA) and an Agent-driven Reasoning and Coding (ARC) framework, GLM-5 substantially enhances long-context modeling and autonomous software engineering proficiency. Experimental results demonstrate that GLM-5 achieves state-of-the-art performance on mainstream open-source benchmarks and significantly outperforms existing models in real-world end-to-end software engineering tasks, validating its superior practical coding capabilities.

Technology Category

Application Category

📝 Abstract
We present GLM-5, a next-generation foundation model designed to transition the paradigm of vibe coding to agentic engineering. Building upon the agentic, reasoning, and coding (ARC) capabilities of its predecessor, GLM-5 adopts DSA to significantly reduce training and inference costs while maintaining long-context fidelity. To advance model alignment and autonomy, we implement a new asynchronous reinforcement learning infrastructure that drastically improves post-training efficiency by decoupling generation from training. Furthermore, we propose novel asynchronous agent RL algorithms that further improve RL quality, enabling the model to learn from complex, long-horizon interactions more effectively. Through these innovations, GLM-5 achieves state-of-the-art performance on major open benchmarks. Most critically, GLM-5 demonstrates unprecedented capability in real-world coding tasks, surpassing previous baselines in handling end-to-end software engineering challenges. Code, models, and more information are available at https://github.com/zai-org/GLM-5.
Problem

Research questions and friction points this paper is trying to address.

vibe coding
agentic engineering
foundation model
software engineering
autonomous agents
Innovation

Methods, ideas, or system contributions that make the work stand out.

asynchronous reinforcement learning
DSA
agentic engineering
long-context modeling
foundation model
🔎 Similar Papers
No similar papers found.
Aohan Zeng
Aohan Zeng
Tsinghua University
Large Language ModelsNatural Language Processing
X
Xin Lv
Zhipu AI & Tsinghua University
Zhenyu Hou
Zhenyu Hou
Tsinghua University
Language model reasoningGraph neural networks
Zhengxiao Du
Zhengxiao Du
清华大学
machine learningpretrained language modelknowledge graph
Qinkai Zheng
Qinkai Zheng
Tsinghua University
Code GenerationGraph Machine LearningAdversarial Robustness
B
Bin Chen
Zhipu AI & Tsinghua University
Da Yin
Da Yin
Meta FAIR
Natural Language Processing
C
Chendi Ge
Zhipu AI & Tsinghua University
C
Chengxing Xie
Zhipu AI & Tsinghua University
Cunxiang Wang
Cunxiang Wang
Tsinghua University; ZhipuAI
Large Language ModelsLLM EvaluationLLM Post-training
G
Gengzheng Pan
Zhipu AI & Tsinghua University
H
Hao Zeng
Zhipu AI & Tsinghua University
H
Haoke Zhang
Zhipu AI & Tsinghua University
Haoran Wang
Haoran Wang
Tsinghua University
Machine Learning‬NLPAgentLLM Post-Training
H
Huilong Chen
Zhipu AI & Tsinghua University
Jiajie Zhang
Jiajie Zhang
Tsinghua University
Large Language ModelNatural Language Processing
J
Jian Jiao
Zhipu AI & Tsinghua University
J
Jiaqi Guo
Zhipu AI & Tsinghua University
J
Jingsen Wang
Zhipu AI & Tsinghua University
J
Jingzhao Du
Zhipu AI & Tsinghua University
J
Jinzhu Wu
Zhipu AI & Tsinghua University
K
Kedong Wang
Zhipu AI & Tsinghua University
L
Lei Li
Zhipu AI & Tsinghua University