π€ AI Summary
To bridge the gap between static LLM pretraining and dynamic agent execution, this paper proposes KAT-Devβa deployable intelligent coding agent framework. Methodologically: (1) it introduces a multi-stage curriculum training pipeline, incorporating reflection-augmented mid-phase training; (2) it proposes multi-ground-truth reward modeling and error-masked supervised fine-tuning to enhance RL efficiency and robustness; and (3) it adopts tree-structured trajectory training to align with production-grade IDE environments. Contributions include: the first framework jointly optimizing long-context reasoning, instruction alignment, and reliable tool invocation. The KAT-Dev 32B model supports 20 programming languages and significantly outperforms baselines in realistic development scenarios. All code, datasets, and models are publicly released, enabling industrial deployment of intelligent programming agents.
π Abstract
Recent advances in large language models (LLMs) have enabled progress in agentic coding, where models autonomously reason, plan, and act within interactive software development workflows. However, bridging the gap between static text-based training and dynamic real-world agentic execution remains a core challenge. In this technical report, we present KAT-Coder, a large-scale agentic code model trained through a multi-stage curriculum encompassing Mid-Term Training, Supervised Fine-Tuning (SFT), Reinforcement Fine-Tuning (RFT), and Reinforcement-to-Deployment Adaptation. The Mid-Term stage enhances reasoning, planning, and reflection capabilities through a corpus of real software engineering data and synthetic agentic interactions. The SFT stage constructs a million-sample dataset balancing twenty programming languages, ten development contexts, and ten task archetypes. The RFT stage introduces a novel multi-ground-truth reward formulation for stable and sample-efficient policy optimization. Finally, the Reinforcement-to-Deployment phase adapts the model to production-grade IDE environments using Error-Masked SFT and Tree-Structured Trajectory Training. In summary, these stages enable KAT-Coder to achieve robust tool-use reliability, instruction alignment, and long-context reasoning, forming a deployable foundation for real-world intelligent coding agents. Our KAT series 32B model, KAT-Dev, has been open-sourced on https://huggingface.co/Kwaipilot/KAT-Dev.