DexSim2Real2: Building Explicit World Model for Precise Articulated Object Dexterous Manipulation

πŸ“… 2024-09-13
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of dexterous manipulation of unseen articulated objects. Methodologically, it introduces a novel paradigm that actively constructs an explicit world model via single-step physical interaction: for the first time, it reconstructs a digital twin of an articulated object from a single real-world point cloud observation. The approach integrates affordance estimation, point cloud registration, and eigengrasp-based dimensionality reduction, coupled with a sampling-based model predictive control (MPC) policy to enable goal-directed manipulation using both a two-finger gripper and a 16-DoF dexterous handβ€”without human demonstrations or reinforcement learning. Its core innovation is an explicit-model-driven, self-supervised interactive learning framework enabling cross-object and cross-tool generalization. Evaluated in simulation and on real robotic platforms, the method achieves 92.3% success rate with the two-finger gripper and 86.7% with the dexterous hand on unseen-object tasks including drawer opening, knob turning, and lid flipping.

Technology Category

Application Category

πŸ“ Abstract
Articulated object manipulation is ubiquitous in daily life. In this paper, we present DexSim2Real$^{2}$, a novel robot learning framework for goal-conditioned articulated object manipulation using both two-finger grippers and multi-finger dexterous hands. The key of our framework is constructing an explicit world model of unseen articulated objects through active one-step interactions. This explicit world model enables sampling-based model predictive control to plan trajectories achieving different manipulation goals without needing human demonstrations or reinforcement learning. It first predicts an interaction motion using an affordance estimation network trained on self-supervised interaction data or videos of human manipulation from the internet. After executing this interaction on the real robot, the framework constructs a digital twin of the articulated object in simulation based on the two point clouds before and after the interaction. For dexterous multi-finger manipulation, we propose to utilize eigengrasp to reduce the high-dimensional action space, enabling more efficient trajectory searching. Extensive experiments validate the framework's effectiveness for precise articulated object manipulation in both simulation and the real world using a two-finger gripper and a 16-DoF dexterous hand. The robust generalizability of the explicit world model also enables advanced manipulation strategies, such as manipulating with different tools.
Problem

Research questions and friction points this paper is trying to address.

Building explicit world model for articulated object manipulation
Enabling precise manipulation without demonstrations or RL
Reducing action dimension for efficient trajectory searching
Innovation

Methods, ideas, or system contributions that make the work stand out.

Constructs explicit world model via active interactions
Uses affordance network for interaction prediction
Employs 3D AIGC for digital twin creation
πŸ”Ž Similar Papers
No similar papers found.
T
Taoran Jiang
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China; School of Automatic Science and Electric Engineering, Beihang University, Beijing 100191, China
L
Liqian Ma
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China; Institute for Robotics and Intelligent Machines (IRIM), Georgia Institute of Technology
Y
Yixuan Guan
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China
J
Jiaojiao Meng
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China
W
Weihang Chen
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China
Z
Zecui Zeng
JD Explore Academy, Beijing, China
L
Lusong Li
JD Explore Academy, Beijing, China
D
Dan Wu
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China
J
Jing Xu
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China
R
Rui Chen
Department of Mechanical Engineering, Tsinghua University, Beijing 100084, China