Learning Generalizable Hand-Object Tracking from Synthetic Demonstrations

📅 2025-12-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the generalization bottleneck in dexterous manipulation caused by scarce real-world human demonstration data, this paper proposes the first purely synthetic-data-driven hand-object tracking and control framework. Methodologically, we design a Hand-Object Planner (HOP) to generate diverse, physically plausible trajectories, and integrate reinforcement learning with interactive imitation learning to build a state-conditioned Hand-Object Tracker (HOT), enabling co-optimization of HOP and HOT. Our contributions are threefold: (1) complete elimination of real demonstrations—long-horizon, complex tasks (e.g., hand retargeting and object rearrangement) are learned solely from synthetic data; (2) zero-shot transfer across unseen object geometries and multi-configurational dexterous hands; and (3) strong cross-domain generalization empirically validated on multiple sim-to-real hand-object systems, significantly alleviating data dependency constraints.

Technology Category

Application Category

📝 Abstract
We present a system for learning generalizable hand-object tracking controllers purely from synthetic data, without requiring any human demonstrations. Our approach makes two key contributions: (1) HOP, a Hand-Object Planner, which can synthesize diverse hand-object trajectories; and (2) HOT, a Hand-Object Tracker that bridges synthetic-to-physical transfer through reinforcement learning and interaction imitation learning, delivering a generalizable controller conditioned on target hand-object states. Our method extends to diverse object shapes and hand morphologies. Through extensive evaluations, we show that our approach enables dexterous hands to track challenging, long-horizon sequences including object re-arrangement and agile in-hand reorientation. These results represent a significant step toward scalable foundation controllers for manipulation that can learn entirely from synthetic data, breaking the data bottleneck that has long constrained progress in dexterous manipulation.
Problem

Research questions and friction points this paper is trying to address.

Develops hand-object tracking controllers from synthetic data
Generalizes across diverse object shapes and hand morphologies
Enables dexterous manipulation without human demonstrations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Synthetic data trains hand-object tracking controllers
HOP planner generates diverse hand-object trajectories
HOT tracker uses reinforcement and imitation learning
🔎 Similar Papers
Yinhuai Wang
Yinhuai Wang
Hong Kong University of Science and Technology
Computer VisionRoboticsEmbodied AI
R
Runyi Yu
HKUST, China
H
Hok Wai Tsui
HKUST, China
X
Xiaoyi Lin
Wuhan University, China
H
Hui Zhang
ETH Zurich, Switzerland
Q
Qihan Zhao
HKUST, China
Ke Fan
Ke Fan
Fudan University
Machine LearningDeep Learning
M
Miao Li
Wuhan University, China
J
Jie Song
HKUST (Guangzhou), China
J
Jingbo Wang
Shanghai AI Lab, China
Qifeng Chen
Qifeng Chen
HKUST
Computational PhotographyImage SynthesisGenerative AIAutonomous DrivingEmbodied AI
Ping Tan
Ping Tan
Hong Kong University of Science and Technology (HKUST)
Computer VisionComputer Graphics