🤖 AI Summary
This work addresses the challenge that small-parameter models (e.g., 32B) underperform large-scale models (e.g., GPT-OSS 120B, DeepSeek v3.1) on complex reasoning tasks. We propose a parameter-efficient inference system built upon the Qwen2.5 architecture and trained exclusively on public data, integrating six core techniques: long-chain reasoning supervised fine-tuning, verifiable reward-based reinforcement learning, agent-driven pre-inference planning, test-time scaling, speculative decoding, and inference hardware co-optimization. Our method achieves state-of-the-art performance among open-source models on mathematical, coding, and scientific reasoning benchmarks—including MATH, GPQA, and LiveCodeBench—while sustaining >2000 tokens/sec throughput per request and enabling lightweight deployment. The approach significantly enhances the reasoning efficiency, cost-effectiveness, and scalability of compact language models.
📝 Abstract
K2-Think is a reasoning system that achieves state-of-the-art performance with a 32B parameter model, matching or surpassing much larger models like GPT-OSS 120B and DeepSeek v3.1. Built on the Qwen2.5 base model, our system shows that smaller models can compete at the highest levels by combining advanced post-training and test-time computation techniques. The approach is based on six key technical pillars: Long Chain-of-thought Supervised Finetuning, Reinforcement Learning with Verifiable Rewards (RLVR), Agentic planning prior to reasoning, Test-time Scaling, Speculative Decoding, and Inference-optimized Hardware, all using publicly available open-source datasets. K2-Think excels in mathematical reasoning, achieving state-of-the-art scores on public benchmarks for open-source models, while also performing strongly in other areas such as Code and Science. Our results confirm that a more parameter-efficient model like K2-Think 32B can compete with state-of-the-art systems through an integrated post-training recipe that includes long chain-of-thought training and strategic inference-time enhancements, making open-source reasoning systems more accessible and affordable. K2-Think is freely available at k2think.ai, offering best-in-class inference speeds of over 2,000 tokens per second per request via the Cerebras Wafer-Scale Engine.