🤖 AI Summary
This work addresses the dual challenges of obstacle-aware trajectory generation and liquid-slosh suppression during vision-only VR teleoperation of a UR5e robotic arm—specifically when manipulating a glass containing liquid. We propose a general shared-control framework that integrates task-oriented multi-objective optimization with real-time motion planning. The method directly maps non-haptic VR controller inputs to joint-space trajectories satisfying collision avoidance, dynamic feasibility, and liquid-slosh suppression constraints—modeled via an equivalent single-pendulum approximation. Crucially, our approach enables online embedding and coordinated regulation of dynamic task constraints without requiring haptic feedback. Experimental evaluation demonstrates an average planning latency of only 13 ms, achieving high operational stability while significantly improving teleoperation accuracy and generalization capability for complex physical interaction tasks, such as steady-state pouring.
📝 Abstract
Teleoperation with non-haptic VR controllers deprives human operators of critical motion feedback. We address this by embedding a multi-objective optimization problem that converts user input into collision-free UR5e joint trajectories while actively suppressing liquid slosh in a glass. The controller maintains 13 ms average planning latency, confirming real-time performance and motivating the augmentation of this teleoperation approach to further objectives.