AgentFrontier: Expanding the Capability Frontier of LLM Agents with ZPD-Guided Data Synthesis

📅 2025-10-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of extending the reasoning capabilities of large language model (LLM) agents beyond current limits. Method: We propose AgentFrontier, a framework grounded in Vygotsky’s Zone of Proximal Development (ZPD) theory, which introduces a dynamic ZPD-based evaluation benchmark for automated identification, synthesis, and assessment of boundary-pushing reasoning tasks. Our approach integrates ZPD-guided, multi-domain high-quality data generation with continual pretraining and targeted post-training—balancing knowledge density and reasoning depth. Contribution/Results: We present the first systematic modeling of ZPD as an operational mechanism for LLM agent capability evolution and design a scalable, data-driven training paradigm accordingly. The resulting AgentFrontier-30B-A3B model achieves state-of-the-art performance on demanding reasoning benchmarks—including Humanity’s Last Exam—outperforming leading proprietary agent models, thereby validating the efficacy of ZPD-driven training for enhancing complex reasoning.

Technology Category

Application Category

📝 Abstract
Training large language model agents on tasks at the frontier of their capabilities is key to unlocking advanced reasoning. We introduce a data synthesis approach inspired by the educational theory of the Zone of Proximal Development (ZPD), which defines this frontier as tasks an LLM cannot solve alone but can master with guidance. To operationalize this, we present the AgentFrontier Engine, an automated pipeline that synthesizes high-quality, multidisciplinary data situated precisely within the LLM's ZPD. This engine supports both continued pre-training with knowledge-intensive data and targeted post-training on complex reasoning tasks. From the same framework, we derive the ZPD Exam, a dynamic and automated benchmark designed to evaluate agent capabilities on these frontier tasks. We train AgentFrontier-30B-A3B model on our synthesized data, which achieves state-of-the-art results on demanding benchmarks like Humanity's Last Exam, even surpassing some leading proprietary agents. Our work demonstrates that a ZPD-guided approach to data synthesis offers a scalable and effective path toward building more capable LLM agents.
Problem

Research questions and friction points this paper is trying to address.

Expanding LLM agent capabilities through ZPD-guided data synthesis
Automating creation of high-quality training data for frontier tasks
Developing dynamic benchmark to evaluate agent reasoning performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

ZPD-guided data synthesis for LLM agents
Automated pipeline generates high-quality ZPD data
Combines pre-training and post-training on frontier tasks
🔎 Similar Papers
No similar papers found.