Integrating Model-based Control and RL for Sim2Real Transfer of Tight Insertion Policies

📅 2025-05-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Achieving sub-millimeter (<1 mm) precision insertion under Sim2Real transfer remains challenging due to modeling inaccuracies and sensorimotor discrepancies. Method: We propose a potential-field-guided residual reinforcement learning (RL) framework. A SE(3)-aware potential-field controller provides stable, interpretable priors, while a residual RL network—trained end-to-end in simulation—refines actions with minimal deviation. We incorporate dual curriculum learning (observation noise and action magnitude), sparse reward training, and vision-based SE(3) target tracking. Contribution/Results: This work is the first to tightly integrate physically grounded, interpretable potential-field control with data-efficient residual RL, enabling zero-shot, real-world deployment without fine-tuning on physical hardware. Experiments demonstrate superior performance over pure RL and existing hybrid approaches across multiple objects and operating conditions, achieving real-time, robust sub-millimeter insertion accuracy.

Technology Category

Application Category

📝 Abstract
Object insertion under tight tolerances ($<hspace{-.02in} 1mm$) is an important but challenging assembly task as even small errors can result in undesirable contacts. Recent efforts focused on Reinforcement Learning (RL), which often depends on careful definition of dense reward functions. This work proposes an effective strategy for such tasks that integrates traditional model-based control with RL to achieve improved insertion accuracy. The policy is trained exclusively in simulation and is zero-shot transferred to the real system. It employs a potential field-based controller to acquire a model-based policy for inserting a plug into a socket given full observability in simulation. This policy is then integrated with residual RL, which is trained in simulation given only a sparse, goal-reaching reward. A curriculum scheme over observation noise and action magnitude is used for training the residual RL policy. Both policy components use as input the SE(3) poses of both the plug and the socket and return the plug's SE(3) pose transform, which is executed by a robotic arm using a controller. The integrated policy is deployed on the real system without further training or fine-tuning, given a visual SE(3) object tracker. The proposed solution and alternatives are evaluated across a variety of objects and conditions in simulation and reality. The proposed approach outperforms recent RL-based methods in this domain and prior efforts with hybrid policies. Ablations highlight the impact of each component of the approach.
Problem

Research questions and friction points this paper is trying to address.

Develops hybrid model-based control and RL for precise tight-tolerance insertion tasks
Enables zero-shot Sim2Real transfer without real-world training or fine-tuning
Improves insertion accuracy over pure RL and prior hybrid methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates model-based control with RL
Uses potential field-based controller in simulation
Zero-shot transfer to real system
🔎 Similar Papers
No similar papers found.