Focused Blind Switching Manipulation Based on Constrained and Regional Touch States of Multi-Fingered Hand Using Deep Learning

📅 2025-03-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Multifinger dexterous hands struggle to achieve high-degree-of-freedom coordinated motion switching under blind conditions—i.e., without visual feedback—especially when responding to subtle tactile variations. Method: This paper proposes a real-time blind grasp pose adjustment strategy integrating tactile constraints and regional attention mechanisms. We design a tactile state constraint loss and a dynamic modality attention gating mechanism, building an AE-LSTM hybrid architecture: an autoencoder compresses whole-hand tactile signals, while an LSTM models the temporal tactile–action mapping; additionally, region-wise tactile state encoding enables subtask-driven motion transitions. Contribution/Results: To our knowledge, this is the first method enabling blind, dynamic motion-mode transfer based on whole-hand regional tactile perception. Evaluated on a physical robot platform across diverse bottle-cap opening tasks, it achieves state-of-the-art success rates. The model autonomously discriminates subtasks (e.g., sliding vs. unscrewing) and adaptively focuses on task-critical tactile sensor modalities.

Technology Category

Application Category

📝 Abstract
To achieve a desired grasping posture (including object position and orientation), multi-finger motions need to be conducted according to the the current touch state. Specifically, when subtle changes happen during correcting the object state, not only proprioception but also tactile information from the entire hand can be beneficial. However, switching motions with high-DOFs of multiple fingers and abundant tactile information is still challenging. In this study, we propose a loss function with constraints of touch states and an attention mechanism for focusing on important modalities depending on the touch states. The policy model is AE-LSTM which consists of Autoencoder (AE) which compresses abundant tactile information and Long Short-Term Memory (LSTM) which switches the motion depending on the touch states. Motion for cap-opening was chosen as a target task which consists of subtasks of sliding an object and opening its cap. As a result, the proposed method achieved the best success rates with a variety of objects for real time cap-opening manipulation. Furthermore, we could confirm that the proposed model acquired the features of each subtask and attention on specific modalities.
Problem

Research questions and friction points this paper is trying to address.

Achieving precise multi-finger grasping posture using touch state feedback.
Handling high-DOF finger motions with abundant tactile information.
Real-time cap-opening manipulation with improved success rates.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deep learning for multi-finger motion control
AE-LSTM model with tactile information compression
Attention mechanism focusing on key touch states
🔎 Similar Papers
No similar papers found.
S
S. Funabashi
Waseda University, Okubo 3-4-1, Shinjuku, Tokyo 169-8555, Japan
A
Atsumu Hiramoto
Waseda University, Okubo 3-4-1, Shinjuku, Tokyo 169-8555, Japan
Naoya Chiba
Naoya Chiba
D3 Center, Osaka University
3D Measurement3D Data ProcessingRobot VisionRobotics
Alexander Schmitz
Alexander Schmitz
Associate Professor at Waseda University
Artificial IntelligenceRobotics
S
Shardul Kulkarni
Waseda University, Okubo 3-4-1, Shinjuku, Tokyo 169-8555, Japan
Tetsuya Ogata
Tetsuya Ogata
Professor, Waseda University / Joint-appointed Fellow, AIST / Visiting Professor, NII
Deep Predictive LearningPhysical AIDevelopmental Robotics