🤖 AI Summary
This work proposes a closed-loop digital twin framework to address the challenges of accuracy and safety in robotic grasping under dynamically occluded visual conditions. The framework integrates an offline 3D asset library with online state synchronization, leveraging VGGT to rapidly reconstruct 3D models from RGB images. By combining point cloud segmentation with color-based ICP registration, it achieves real-time alignment between the physical and simulated environments. This synchronized digital twin enables robust grasp planning and safe execution. Experimental results demonstrate that the proposed method significantly improves both grasping success rates and motion safety in dynamic occlusion scenarios, validating the effectiveness and novelty of the digital twin mechanism for real-world robotic tasks.
📝 Abstract
Accurate and safe grasping under dynamic and visually occluded conditions remains a core challenge in real-world robotic manipulation. We present SyncTwin, a digital twin framework that unifies fast 3D scene reconstruction and real-to-sim synchronization for robust and safety-aware grasping in such environments. In the offline stage, we employ VGGT to rapidly reconstruct object-level 3D assets from RGB images, forming a reusable geometry library for simulation. During execution, SyncTwin continuously synchronizes the digital twin by tracking real-world object states via point cloud segmentation updates and aligning them through colored-ICP registration. The updated twin enables motion planners to compute collision-free and dynamically feasible trajectories in simulation, which are safely executed on the real robot through a closed real-to-sim-to-real loop. Experiments in dynamic and occluded scenes show that SyncTwin improves grasp accuracy and motion safety, demonstrating the effectiveness of digital-twin synchronization for real-world robotic execution.