AnyTask: an Automated Task and Data Generation Framework for Advancing Sim-to-Real Policy Learning

📅 2025-12-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the scarcity of real-world interaction data and the high cost of manually designing simulation tasks in general-purpose robotic learning, this paper introduces the first fully automated task and dataset generation framework. The framework integrates GPU-accelerated parallel physics simulation with multimodal foundation models (VLMs and LLMs) to enable end-to-end automation of task design, scene generation, and expert demonstration synthesis. We propose the ViPR family of agents—ViPR, ViPR-Eureka, and ViPR-RL—which incorporate VLM-in-the-loop planning, LLM-guided contact sampling, and hybrid policy learning under sparse rewards, thereby significantly reducing reliance on human intervention in conventional sim-to-real pipelines. Critically, policies trained on generated data deploy directly to physical robots without fine-tuning. Evaluated on pick-and-place, drawer opening, contact-intensive pushing, and long-horizon manipulation tasks, the approach achieves an average success rate of 44%.

Technology Category

Application Category

📝 Abstract
Generalist robot learning remains constrained by data: large-scale, diverse, and high-quality interaction data are expensive to collect in the real world. While simulation has become a promising way for scaling up data collection, the related tasks, including simulation task design, task-aware scene generation, expert demonstration synthesis, and sim-to-real transfer, still demand substantial human effort. We present AnyTask, an automated framework that pairs massively parallel GPU simulation with foundation models to design diverse manipulation tasks and synthesize robot data. We introduce three AnyTask agents for generating expert demonstrations aiming to solve as many tasks as possible: 1) ViPR, a novel task and motion planning agent with VLM-in-the-loop Parallel Refinement; 2) ViPR-Eureka, a reinforcement learning agent with generated dense rewards and LLM-guided contact sampling; 3) ViPR-RL, a hybrid planning and learning approach that jointly produces high-quality demonstrations with only sparse rewards. We train behavior cloning policies on generated data, validate them in simulation, and deploy them directly on real robot hardware. The policies generalize to novel object poses, achieving 44% average success across a suite of real-world pick-and-place, drawer opening, contact-rich pushing, and long-horizon manipulation tasks. Our project website is at https://anytask.rai-inst.com .
Problem

Research questions and friction points this paper is trying to address.

Automates robot task design and data generation
Reduces human effort in sim-to-real transfer
Enables generalization to novel object poses
Innovation

Methods, ideas, or system contributions that make the work stand out.

Automated task generation using foundation models and GPU simulation
Three specialized agents for expert demonstration synthesis
Behavior cloning policies trained on generated data for sim-to-real transfer
🔎 Similar Papers
No similar papers found.
R
Ran Gong
Robotics and AI Institute, Boston, MA, USA
X
Xiaohan Zhang
Robotics and AI Institute, Boston, MA, USA
Jinghuan Shang
Jinghuan Shang
Research Scientist, Robotics and AI Institute
Representation LearningRoboticsComputer Vision
Maria Vittoria Minniti
Maria Vittoria Minniti
RAI Institute
RoboticsControlArtificial IntelligenceMachine Learning
J
Jigarkumar Patel
Robotics and AI Institute, Boston, MA, USA
V
Valerio Pepe
Robotics and AI Institute, Boston, MA, USA
R
Riedana Yan
Robotics and AI Institute, Boston, MA, USA
A
Ahmet Gundogdu
Robotics and AI Institute, Boston, MA, USA
Ivan Kapelyukh
Ivan Kapelyukh
Imperial College London
machine learningrobot learninggraph neural networksreinforcement learningrepresentation learning
A
Ali Abbas
Robotics and AI Institute, Boston, MA, USA
X
Xiaoqiang Yan
Robotics and AI Institute, Boston, MA, USA
H
Harsh Patel
Robotics and AI Institute, Boston, MA, USA
Laura Herlant
Laura Herlant
Director of Robot Perception, Robotics and AI Institute
Karl Schmeckpeper
Karl Schmeckpeper
The AI Institute
RoboticsMachine LearningComputer Vision