🤖 AI Summary
Large language models (LLMs) exhibit inherent limitations in real-world spatial reasoning and question answering—particularly inaccurate spatial data retrieval and weak geometric relation inference. To address these challenges, we propose an LLM-guided spatially aware generation paradigm featuring a novel multi-objective weighted ranking mechanism that jointly integrates sparse spatial retrieval (exact queries over spatial databases) and dense semantic retrieval (embedding-based similarity matching). This approach synergistically leverages structured spatial constraints and unstructured semantic information to guide LLMs in generating accurate, interpretable spatial answers. Extensive experiments on a real-world tourism-oriented spatial QA dataset demonstrate significant improvements in answer accuracy. Our method effectively bridges the critical capability gap between general-purpose LLMs and spatial intelligence, establishing a new paradigm for geographic AI and spatial foundation models.
📝 Abstract
Spatial reasoning remains a challenge for Large Language Models (LLMs), which struggle with spatial data retrieval and reasoning. We propose Spatial Retrieval-Augmented Generation (Spatial-RAG), a framework that extends RAG to spatial tasks by integrating sparse spatial retrieval (spatial databases) and dense semantic retrieval (LLM-based similarity). A multi-objective ranking strategy balances spatial constraints and semantic relevance, while an LLM-guided generator ensures coherent responses. Experiments on a real-world tourism dataset show that Spatial-RAG significantly improves spatial question answering, bridging the gap between LLMs and spatial intelligence.