GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models

πŸ“… 2025-08-08
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of developing a parameter-efficient, high-performance open-source Mixture-of-Experts (MoE) language modelβ€”GLM-4.5β€”that excels in agent behavior, complex reasoning, and code generation. We propose a dual-mode hybrid inference mechanism (Chain-of-Thought reasoning vs. direct response) and a multi-stage MoE expert iterative training framework, enabling high performance with sparse parameter activation. The model is trained on 23T tokens via multi-phase pretraining, reinforcement learning, and fine-grained post-training. Evaluated on TAU-Bench (70.1%), AIME 2024 (91.0%), and SWE-bench Verified (64.2%), GLM-4.5 ranks third overall and second on agent-centric tasks. To foster research in lightweight large language models, we publicly release both full-scale and compact variants of GLM-4.5.

Technology Category

Application Category

πŸ“ Abstract
We present GLM-4.5, an open-source Mixture-of-Experts (MoE) large language model with 355B total parameters and 32B activated parameters, featuring a hybrid reasoning method that supports both thinking and direct response modes. Through multi-stage training on 23T tokens and comprehensive post-training with expert model iteration and reinforcement learning, GLM-4.5 achieves strong performance across agentic, reasoning, and coding (ARC) tasks, scoring 70.1% on TAU-Bench, 91.0% on AIME 24, and 64.2% on SWE-bench Verified. With much fewer parameters than several competitors, GLM-4.5 ranks 3rd overall among all evaluated models and 2nd on agentic benchmarks. We release both GLM-4.5 (355B parameters) and a compact version, GLM-4.5-Air (106B parameters), to advance research in reasoning and agentic AI systems. Code, models, and more information are available at https://github.com/zai-org/GLM-4.5.
Problem

Research questions and friction points this paper is trying to address.

Develops hybrid reasoning for thinking and response modes
Enhances performance in agentic, reasoning, and coding tasks
Provides open-source models for AI reasoning research
Innovation

Methods, ideas, or system contributions that make the work stand out.

Mixture-of-Experts large language model
Hybrid reasoning with dual response modes
Multi-stage training and expert iteration
πŸ”Ž Similar Papers
No similar papers found.
Aohan Zeng
Aohan Zeng
Tsinghua University
Large Language ModelsNatural Language Processing
X
Xin Lv
Zhipu AI & Tsinghua University
Qinkai Zheng
Qinkai Zheng
Tsinghua University
Code GenerationGraph Machine LearningAdversarial Robustness
Zhenyu Hou
Zhenyu Hou
Tsinghua University
Language model reasoningGraph neural networks
B
Bin Chen
Zhipu AI & Tsinghua University
C
Chengxing Xie
Zhipu AI & Tsinghua University
Cunxiang Wang
Cunxiang Wang
Tsinghua University; ZhipuAI
Large Language ModelsLLM EvaluationLLM Post-training
Da Yin
Da Yin
Meta FAIR
Natural Language Processing
H
Hao Zeng
Zhipu AI & Tsinghua University
J
Jiajie Zhang
Zhipu AI & Tsinghua University
K
Kedong Wang
Zhipu AI & Tsinghua University
Lucen Zhong
Lucen Zhong
Zhipu AI
LLMFunction CallingAgentic Tool Using
Mingdao Liu
Mingdao Liu
PhD Student, Tsinghua University
Natural Language Processing
R
Rui Lu
Zhipu AI & Tsinghua University
S
Shulin Cao
Zhipu AI & Tsinghua University
X
Xiaohan Zhang
Zhipu AI & Tsinghua University
X
Xuancheng Huang
Zhipu AI & Tsinghua University
Yao Wei
Yao Wei
Zhejiang University
Foundation ModelEmbodied AILarge Language Model
Y
Yean Cheng
Zhipu AI & Tsinghua University
Y
Yifan An
Zhipu AI & Tsinghua University
Yilin Niu
Yilin Niu
Tsinghua University
Natural Language Processing
Y
Yuanhao Wen
Zhipu AI & Tsinghua University
Yushi Bai
Yushi Bai
Tsinghua University
Large Language ModelsMachine LearningKnowledge GraphAlgorithmic Game Theory