URDF-Anything+: Autoregressive Articulated 3D Models Generation for Physical Simulation

📅 2026-03-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Jointly reconstructing the part geometry and kinematic structure of articulated objects from visual inputs remains a significant challenge. This work proposes the first end-to-end autoregressive framework that directly generates complete, executable URDF models from images and object-level 3D cues, sequentially predicting part geometries and joint parameters while automatically inferring the underlying kinematic structure. Notably, this approach achieves single-stage generation of full URDF specifications—a capability not demonstrated by prior methods—and enables the Real-Follow-Sim paradigm, allowing policies trained purely in simulation to be transferred to real robots without any fine-tuning. Evaluated on large-scale benchmarks and real-world tasks, the method substantially outperforms existing approaches in terms of geometric reconstruction fidelity, joint parameter accuracy, and physical executability.

Technology Category

Application Category

📝 Abstract
Articulated objects are fundamental for robotics, simulation of physics, and interactive virtual environments. However, reconstructing them from visual input remains challenging, as it requires jointly inferring both part geometry and kinematic structure. We present, an end-to-end autoregressive framework that directly generates executable articulated object models from visual observations. Given image and object-level 3D cues, our method sequentially produces part geometries and their associated joint parameters, resulting in complete URDF models without reliance on multi-stage pipelines. The generation proceeds until the model determines that all parts have been produced, automatically inferring complete geometry and kinematics. Building on this capability, we enable a new Real-Follow-Sim paradigm, where high-fidelity digital twins constructed from visual observations allow policies trained and tested purely in simulation to transfer to real robots without online adaptation. Experiments on large-scale articulated object benchmarks and real-world robotic tasks demonstrate that outperforms prior methods in geometric reconstruction quality, joint parameter accuracy, and physical executability.
Problem

Research questions and friction points this paper is trying to address.

articulated objects
3D reconstruction
kinematic structure
visual input
geometric inference
Innovation

Methods, ideas, or system contributions that make the work stand out.

autoregressive generation
articulated 3D models
URDF
physical simulation
sim-to-real transfer
🔎 Similar Papers
No similar papers found.