🤖 AI Summary
Current robotic manipulation policies rely on global or dense visual representations, which conflate task-relevant and irrelevant information, resulting in poor generalization under distribution shifts—such as lighting variations, texture changes, and occluding distractors. This work proposes object-centric representations (OCR) as a structured alternative, explicitly decomposing scenes into discrete entities and embedding task-appropriate inductive biases for manipulation. We present the first systematic evaluation demonstrating that OCR significantly outperforms global (CNN/Transformer) and dense representations—even without task-specific pretraining. Through comparative experiments across multiple encoder architectures, OCR achieves an average 23.5% improvement in generalization performance on multi-task benchmarks in both simulation and real-world settings. Notably, it maintains high success rates under severe visual distractors and cross-domain conditions, validating its robustness and scalability for real-world deployment.
📝 Abstract
Visual representations are central to the learning and generalization capabilities of robotic manipulation policies. While existing methods rely on global or dense features, such representations often entangle task-relevant and irrelevant scene information, limiting robustness under distribution shifts. In this work, we investigate object-centric representations (OCR) as a structured alternative that segments visual input into a finished set of entities, introducing inductive biases that align more naturally with manipulation tasks. We benchmark a range of visual encoders-object-centric, global and dense methods-across a suite of simulated and real-world manipulation tasks ranging from simple to complex, and evaluate their generalization under diverse visual conditions including changes in lighting, texture, and the presence of distractors. Our findings reveal that OCR-based policies outperform dense and global representations in generalization settings, even without task-specific pretraining. These insights suggest that OCR is a promising direction for designing visual systems that generalize effectively in dynamic, real-world robotic environments.