GTM: Simulating the World of Tools for AI Agents

📅 2025-12-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high computational cost, low efficiency, and substantial maintenance overhead associated with integrating external tools into LLM-based agents, this paper proposes the Generalist Tool Model (GTM)—a lightweight (1.5B-parameter) sequence model. GTM employs Context-Aware Response Generation (CARG) and a prompt-level data synthesis pipeline to faithfully emulate over 20,000 cross-domain tools. Crucially, GTM enables agent training without invoking real tools, achieving— for the first time—zero-shot cross-domain tool generalization and strong contextual consistency. Experiments demonstrate that GTM’s inference speed exceeds that of real tools by several orders of magnitude, while maintaining comparable output accuracy. Moreover, in reinforcement learning settings, GTM significantly enhances training scalability and deployment flexibility.

Technology Category

Application Category

📝 Abstract
The integration of external tools is pivotal for empowering Large Language Model (LLM) agents with real-world capabilities. However, training these agents through direct, continuous interaction with diverse tools is often prohibitively expensive, slow, and introduces additional development and maintenance overhead. To address this challenge, we introduce the Generalist Tool Model (GTM), a 1.5-billion-parameter model that learns to act as a universal tool simulator. With only prompt-level configuration, GTM accesses tool functionalities along with input arguments and generates outputs that faithfully mimic real tool execution, providing a fast and cost-effective solution that eliminates development overhead. To build GTM, we propose the Context-Aware Response Generation (CARG) pipeline, which synthesizes comprehensive training data covering over 20,000 tools across 300 domains including physics, medicine, robotics, and finance. Through this pipeline, GTM learns to produce not only syntactically correct outputs but also logically coherent and contextually appropriate responses. Experiments demonstrate that GTM produces high-quality outputs with strong consistency and reliability. Besides when used in real reinforcement learning scenarios for agent training, GTM exhibits significantly faster simulation speed compared to real tools while maintaining comparable output quality, along with remarkable generalization and domain adaptability. Our results establish GTM as a foundational component for developing future AI agents, enabling efficient and scalable training of tool-augmented systems.
Problem

Research questions and friction points this paper is trying to address.

Simulates diverse tools to reduce AI agent training costs
Generates realistic tool outputs without direct tool integration
Enables scalable training across multiple domains efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

GTM simulates diverse tools using 1.5B parameters
CARG pipeline synthesizes training data across 300 domains
GTM enables fast, cost-effective agent training via simulation
🔎 Similar Papers
No similar papers found.
Z
Zhenzhen Ren
School of Computer Science, Fudan University, Shanghai, China
X
Xinpeng Zhang
School of Computer Science, Fudan University, Shanghai, China
Z
Zhenxing Qian
School of Computer Science, Fudan University, Shanghai, China
Y
Yan Gao
Zhongguancun Academy, Beijing, China
Y
Yu Shi
Zhongguancun Academy, Beijing, China
Shuxin Zheng
Shuxin Zheng
Deputy Director, Zhongguancun Institute of Artificial Intelligence
General AIGenerative AI
Jiyan He
Jiyan He
University of Science and Technology of China
Machine LearningAI for Science