π€ AI Summary
This work addresses the challenge of enabling robots to learn transferable 6-DoF object manipulation skills from only a few uncalibrated, monocular RGB videos of human demonstrations. We propose a Neural Affordance Function that jointly models geometric, visual, and functional cues, yielding a lightweight and reusable manipulation skill memory. Our method employs a multimodal feature continuous-query mechanism to achieve end-to-end mapping from affordance representations to precise manipulation policies, supporting both coarse-to-fine and cross-task/cross-environment policy transfer. Evaluated on 13 simulated and real-world manipulation tasks, our approach achieves an average success rate improvement of 14.9 percentage points over prior methods, requiring only 2β3 demonstration videos per task. The framework demonstrates strong generalization across objects, environments, and task variants while maintaining computational efficiency and scalability.
π Abstract
We present Actron3D, a framework that enables robots to acquire transferable 6-DoF manipulation skills from just a few monocular, uncalibrated, RGB-only human videos. At its core lies the Neural Affordance Function, a compact object-centric representation that distills actionable cues from diverse uncalibrated videos-geometry, visual appearance, and affordance-into a lightweight neural network, forming a memory bank of manipulation skills. During deployment, we adopt a pipeline that retrieves relevant affordance functions and transfers precise 6-DoF manipulation policies via coarse-to-fine optimization, enabled by continuous queries to the multimodal features encoded in the neural functions. Experiments in both simulation and the real world demonstrate that Actron3D significantly outperforms prior methods, achieving a 14.9 percentage point improvement in average success rate across 13 tasks while requiring only 2-3 demonstration videos per task.