🤖 AI Summary
GelSight Mini optical tactile sensors struggle to accurately reconstruct spatially resolved normal and tangential force distributions from silicone deformation images. Method: We propose a physics-guided deep learning approach that leverages high-fidelity finite element analysis (FEA) to generate pixel-level ground-truth 3D force maps—used as supervision to train a U-Net architecture for direct regression of full-contact-area force fields from raw tactile images. Contribution/Results: The method significantly improves cross-indenter and cross-sensor generalization and enables real-time inference (>30 FPS). Evaluated on a novel open-source dataset, it achieves <8.2% error in normal force prediction and <12.6% in tangential force prediction. To foster reproducibility and benchmarking, we publicly release the source code, FEA simulation models, annotated dataset, and pre-trained models—establishing a new standard for vision-based tactile force perception research.
📝 Abstract
Contact-rich manipulation remains a major challenge in robotics. Optical tactile sensors like GelSight Mini offer a low-cost solution for contact sensing by capturing soft-body deformations of the silicone gel. However, accurately inferring shear and normal force distributions from these gel deformations has yet to be fully addressed. In this work, we propose a machine learning approach using a U-net architecture to predict force distributions directly from the sensor's raw images. Our model, trained on force distributions inferred from Finite Element Analysis (FEA), demonstrates promising accuracy in predicting normal and shear force distributions for the commercially available GelSight Mini sensor. It also shows potential for generalization across indenters, sensors of the same type, and for enabling real-time application. The codebase, dataset and models are open-sourced and available at https://feats-ai.github.io .