π€ AI Summary
This work addresses the challenge of generalizing dexterous manipulation across morphologically diverse robots. Methodologically, it introduces a morphology-agnostic unified representation and policy transfer framework: (1) modeling multi-morphology embodiments and actions as 3D particle sets with displacement vectors to yield state-action representations that preserve control semantics while being morphology-invariant; (2) designing a graph-structured world model trained jointly on simulated data, real human hand demonstrations, and exploratory data from heterogeneous robots; and (3) integrating model-predictive control for cross-hardware deployment. To our knowledge, this is the first approach enabling joint training on human hand and robotic hand data. Evaluated on rigid and deformable object manipulation tasks, the framework demonstrates improved generalization with increasing numbers of training morphologies, significantly outperforming morphology-specific baselines and enabling seamless transfer across robots with vastly different degrees of freedomβfrom low-DOF to high-DOF heterogeneous platforms.
π Abstract
Cross-embodiment learning seeks to build generalist robots that operate across diverse morphologies, but differences in action spaces and kinematics hinder data sharing and policy transfer. This raises a central question: Is there any invariance that allows actions to transfer across embodiments? We conjecture that environment dynamics are embodiment-invariant, and that world models capturing these dynamics can provide a unified interface across embodiments. To learn such a unified world model, the crucial step is to design state and action representations that abstract away embodiment-specific details while preserving control relevance. To this end, we represent different embodiments (e.g., human hands and robot hands) as sets of 3D particles and define actions as particle displacements, creating a shared representation for heterogeneous data and control problems. A graph-based world model is then trained on exploration data from diverse simulated robot hands and real human hands, and integrated with model-based planning for deployment on novel hardware. Experiments on rigid and deformable manipulation tasks reveal three findings: (i) scaling to more training embodiments improves generalization to unseen ones, (ii) co-training on both simulated and real data outperforms training on either alone, and (iii) the learned models enable effective control on robots with varied degrees of freedom. These results establish world models as a promising interface for cross-embodiment dexterous manipulation.