🤖 AI Summary
To address the failure of robotic grasp planning caused by occlusion in unstructured bin-picking scenarios, this paper proposes a real-time, class-agnostic invisible-region segmentation method based on Vision Transformers (ViT). Leveraging global self-attention to model long-range dependencies, the approach end-to-end recovers complete object masks—including occluded and non-visible regions. We introduce novel single-head and dual-head ViT architectures: the former performs invisible-region segmentation alone, while the latter jointly segments both invisible and occluding regions. Furthermore, we construct ViTA-SimData—the first photorealistic synthetic dataset specifically designed for industrial bin-picking scenes. Evaluated on COOCA and KINS benchmarks, our method achieves high-accuracy, low-latency joint segmentation of invisible and occluding regions, maintaining real-time performance while significantly improving robotic grasping robustness.
📝 Abstract
Occlusions in robotic bin picking compromise accurate and reliable grasp planning. We present ViTA-Seg, a class-agnostic Vision Transformer framework for real-time amodal segmentation that leverages global attention to recover complete object masks, including hidden regions. We proposte two architectures: a) Single-Head for amodal mask prediction; b) Dual-Head for amodal and occluded mask prediction. We also introduce ViTA-SimData, a photo-realistic synthetic dataset tailored to industrial bin-picking scenario. Extensive experiments on two amodal benchmarks, COOCA and KINS, demonstrate that ViTA-Seg Dual Head achieves strong amodal and occlusion segmentation accuracy with computational efficiency, enabling robust, real-time robotic manipulation.