A Spatial Relationship Aware Dataset for Robotics

📅 2025-06-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Accurate modeling of fine-grained spatial relationships among objects—such as co-occurrence of visually similar objects and multi-level spatial nesting—is critical for real-world robotic task planning in complex indoor environments; however, existing datasets lack 3D positional annotations and explicit spatial relation labels. To address this gap, we introduce SpotSRD, the first spatial-relation-aware dataset designed for real-robot deployment, comprising nearly 1,000 indoor images captured by Boston Dynamics’ Spot robot, with fine-grained annotations of object attributes, 3D coordinates, and 24 spatial relation types. We develop a custom annotation tool and a scene-graph evaluation framework, systematically exposing spatial modeling bottlenecks across six state-of-the-art models. Furthermore, we propose a spatially aware prompting mechanism that injects structured spatial relations into GPT-4o, significantly improving its accuracy on spatial planning tasks. The dataset, annotation tool, and code are publicly released.

Technology Category

Application Category

📝 Abstract
Robotic task planning in real-world environments requires not only object recognition but also a nuanced understanding of spatial relationships between objects. We present a spatial-relationship-aware dataset of nearly 1,000 robot-acquired indoor images, annotated with object attributes, positions, and detailed spatial relationships. Captured using a Boston Dynamics Spot robot and labelled with a custom annotation tool, the dataset reflects complex scenarios with similar or identical objects and intricate spatial arrangements. We benchmark six state-of-the-art scene-graph generation models on this dataset, analysing their inference speed and relational accuracy. Our results highlight significant differences in model performance and demonstrate that integrating explicit spatial relationships into foundation models, such as ChatGPT 4o, substantially improves their ability to generate executable, spatially-aware plans for robotics. The dataset and annotation tool are publicly available at https://github.com/PengPaulWang/SpatialAwareRobotDataset, supporting further research in spatial reasoning for robotics.
Problem

Research questions and friction points this paper is trying to address.

Enhancing robotic task planning with spatial relationship understanding
Evaluating scene-graph models for spatial reasoning accuracy and speed
Improving foundation models' ability to generate spatially-aware robotic plans
Innovation

Methods, ideas, or system contributions that make the work stand out.

Spatial-relationship-aware dataset with robot-acquired images
Custom annotation tool for detailed spatial relationships
Integration of spatial relationships into foundation models
🔎 Similar Papers
No similar papers found.
P
Peng Wang
Manchester Metropolitan University, Manchester, UK
M
Minh Huy Pham
Independent Researcher, Manchester, UK
Zhihao Guo
Zhihao Guo
Manchester Metropolitan University
3D ReconstructionComputer vision
W
Wei Zhou
Cardiff University, Cardiff, UK