Generative AI for Urban Planning: Synthesizing Satellite Imagery via Diffusion Models

📅 2025-05-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current generative AI models struggle to synthesize high-fidelity satellite imagery at scale for urban planning—particularly in simultaneously satisfying realism, practical utility, and constraints from land use, infrastructure, and natural environments. To address this, we propose the first ControlNet-augmented Stable Diffusion framework for controllable remote sensing image synthesis. Our approach introduces a novel OpenStreetMap–satellite image spatial alignment data paradigm, enabling cross-city generalization and fine-grained semantic control. By integrating multimodal conditional generation with geospatial inputs, the framework supports scenario customization and design exploration. Evaluated across three major U.S. cities, our method achieves significantly lower FID and KID scores than baseline methods. Expert evaluations by urban planners and public participants further confirm that generated imagery surpasses real satellite imagery in perceptual realism, visual diversity, and alignment with user intent.

Technology Category

Application Category

📝 Abstract
Generative AI offers new opportunities for automating urban planning by creating site-specific urban layouts and enabling flexible design exploration. However, existing approaches often struggle to produce realistic and practical designs at scale. Therefore, we adapt a state-of-the-art Stable Diffusion model, extended with ControlNet, to generate high-fidelity satellite imagery conditioned on land use descriptions, infrastructure, and natural environments. To overcome data availability limitations, we spatially link satellite imagery with structured land use and constraint information from OpenStreetMap. Using data from three major U.S. cities, we demonstrate that the proposed diffusion model generates realistic and diverse urban landscapes by varying land-use configurations, road networks, and water bodies, facilitating cross-city learning and design diversity. We also systematically evaluate the impacts of varying language prompts and control imagery on the quality of satellite imagery generation. Our model achieves high FID and KID scores and demonstrates robustness across diverse urban contexts. Qualitative assessments from urban planners and the general public show that generated images align closely with design descriptions and constraints, and are often preferred over real images. This work establishes a benchmark for controlled urban imagery generation and highlights the potential of generative AI as a tool for enhancing planning workflows and public engagement.
Problem

Research questions and friction points this paper is trying to address.

Generating realistic urban layouts using diffusion models
Overcoming data limitations with OpenStreetMap integration
Evaluating AI-generated urban imagery for planning workflows
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adapts Stable Diffusion with ControlNet extension
Links satellite imagery with OpenStreetMap data
Generates high-fidelity urban landscapes via land-use prompts
🔎 Similar Papers
No similar papers found.
Q
Qingyi Wang
Department of Civil and Environment Engineering, Massachusetts Institute of Technology
Yuebing Liang
Yuebing Liang
Massachusetts Institute of Technology, University of Hong Kong
urban computingintelligent transportation
Yunhan Zheng
Yunhan Zheng
MIT
travel behaviormachine learningurban economicstransportation policy
K
Kaiyuan Xu
Department of Systems Engineering, Boston University
J
Jinhua Zhao
Department of Urban Studies and Planning, Massachusetts Institute of Technology
Shenhao Wang
Shenhao Wang
University of Florida; Massachusetts Institute of Technology
Urban AIComputational Social ScienceTravel BehaviorUrban SystemsResilience