Grasp Like Humans: Learning Generalizable Multi-Fingered Grasping from Human Proprioceptive Sensorimotor Integration

📅 2025-09-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of direct tactile-proprioceptive-to-action mapping in dexterous multi-fingered robotic grasping, aiming for robust and generalizable manipulation of both rigid and deformable objects. We propose a unified graph-structured polar-coordinate multimodal representation that explicitly encodes hand morphology variations across platforms. To realize cross-platform perception–action mapping, we design the Tactile-Kinematic Spatio-Temporal Graph Network (TK-STGN), which integrates multidimensional subgraph convolutions with attention-enhanced LSTM modules. Leveraging human hand demonstrations collected via data gloves, our approach combines imitation learning with hybrid force–position control. Extensive experiments on multiple robotic platforms demonstrate significant improvements in grasping success rates for unseen and deformable objects. To the best of our knowledge, this is the first framework enabling proprioceptively grounded, multimodal perception–action transfer and generalization across diverse robotic hands.

Technology Category

Application Category

📝 Abstract
Tactile and kinesthetic perceptions are crucial for human dexterous manipulation, enabling reliable grasping of objects via proprioceptive sensorimotor integration. For robotic hands, even though acquiring such tactile and kinesthetic feedback is feasible, establishing a direct mapping from this sensory feedback to motor actions remains challenging. In this paper, we propose a novel glove-mediated tactile-kinematic perception-prediction framework for grasp skill transfer from human intuitive and natural operation to robotic execution based on imitation learning, and its effectiveness is validated through generalized grasping tasks, including those involving deformable objects. Firstly, we integrate a data glove to capture tactile and kinesthetic data at the joint level. The glove is adaptable for both human and robotic hands, allowing data collection from natural human hand demonstrations across different scenarios. It ensures consistency in the raw data format, enabling evaluation of grasping for both human and robotic hands. Secondly, we establish a unified representation of multi-modal inputs based on graph structures with polar coordinates. We explicitly integrate the morphological differences into the designed representation, enhancing the compatibility across different demonstrators and robotic hands. Furthermore, we introduce the Tactile-Kinesthetic Spatio-Temporal Graph Networks (TK-STGN), which leverage multidimensional subgraph convolutions and attention-based LSTM layers to extract spatio-temporal features from graph inputs to predict node-based states for each hand joint. These predictions are then mapped to final commands through a force-position hybrid mapping.
Problem

Research questions and friction points this paper is trying to address.

Learning robotic grasping from human proprioceptive sensorimotor integration
Establishing direct sensory feedback to motor action mapping
Transferring human grasp skills to robots via imitation learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Glove-mediated tactile-kinesthetic perception-prediction framework
Unified graph representation with polar coordinates for compatibility
TK-STGN network with multidimensional convolutions and attention LSTM
🔎 Similar Papers
No similar papers found.
Ce Guo
Ce Guo
Imperial College London
Reconfigurable computingRisk Management
Xieyuanli Chen
Xieyuanli Chen
Associate Professor, NUDT, China
RoboticsSLAMLocalizationLiDAR PerceptionRobot Learning
Z
Zhiwen Zeng
College of Intelligence Science and Technology, National University of Defense Technology
Zirui Guo
Zirui Guo
Beijing University of Posts and Telecommunications
Contrastive learningGraph representation learningRecommendation
Y
Yihong Li
College of Intelligence Science and Technology, National University of Defense Technology
H
Haoran Xiao
College of Intelligence Science and Technology, National University of Defense Technology
D
Dewen Hu
College of Intelligence Science and Technology, National University of Defense Technology
Huimin Lu
Huimin Lu
National University of Defense Technology
Robot VisionMulti-robot CoordinationRobot SoccerRobot Rescue