UniT: Data Efficient Tactile Representation with Generalization to Unseen Objects

πŸ“… 2024-08-12
πŸ“ˆ Citations: 6
✨ Influential: 1
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of few-shot cross-task zero-shot transfer in tactile learning. We propose UniTβ€”the first method to learn general-purpose tactile representations from a single, simple-object tactile image. UniT leverages a VQGAN-based architecture to construct a compact, disentangled latent representation space for tactile data, requiring no task-specific annotations or additional fine-tuning. The learned representations enable direct zero-shot transfer to diverse downstream tasks, including tactile perception (e.g., in-hand 3D/6D pose estimation and tactile classification) and manipulation policy learning. Compared to existing visual and tactile representation learning approaches, UniT achieves state-of-the-art performance across multiple benchmarks. It is successfully deployed on three real-world robotic manipulation tasks, demonstrating key advantages: plug-and-play usability, extreme data efficiency (single-object training), and task-agnostic generalization.

Technology Category

Application Category

πŸ“ Abstract
UniT is an approach to tactile representation learning, using VQGAN to learn a compact latent space and serve as the tactile representation. It uses tactile images obtained from a single simple object to train the representation with generalizability. This tactile representation can be zero-shot transferred to various downstream tasks, including perception tasks and manipulation policy learning. Our benchmarkings on in-hand 3D pose and 6D pose estimation tasks and a tactile classification task show that UniT outperforms existing visual and tactile representation learning methods. Additionally, UniT's effectiveness in policy learning is demonstrated across three real-world tasks involving diverse manipulated objects and complex robot-object-environment interactions. Through extensive experimentation, UniT is shown to be a simple-to-train, plug-and-play, yet widely effective method for tactile representation learning. For more details, please refer to our open-source repository https://github.com/ZhengtongXu/UniT and the project website https://zhengtongxu.github.io/unit-website/.
Problem

Research questions and friction points this paper is trying to address.

Learning compact tactile representations from minimal object data
Zero-shot transfer to diverse perception and manipulation tasks
Outperforming existing methods in pose estimation and classification
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses VQGAN for compact tactile representation
Trains with single object for generalizability
Zero-shot transfer to diverse downstream tasks
πŸ”Ž Similar Papers
No similar papers found.
Zhengtong Xu
Zhengtong Xu
PhD candidate at Purdue University
Robot Learning
R
Raghava Uppuluri
Elmore Family School of Electrical Engineering, Purdue University, West Lafayette, USA
X
Xinwei Zhang
School of Industrial Engineering, Purdue University, West Lafayette, USA
C
Cael Fitch
School of Mechanical Engineering, Purdue University, West Lafayette, USA
P
Philip Glen Crandall
Department of Food Science, University of Arkansas, Fayetteville, USA
Wan Shou
Wan Shou
University of Arkansas; MIT postdoc; MST Ph.D.
Laser ProcessingMicro/nano ManufacturingFunctional FiberCompositesAI&Robotics
D
Dongyi Wang
Department of Biological and Agricultural Engineering, University of Arkansas, Fayetteville, USA
Yu She
Yu She
Assistant Professor, Purdue University
Robotic ManipulationMechanism DesignTactile SensingRobot Learning