RealDriveSim: A Realistic Multi-Modal Multi-Task Synthetic Dataset for Autonomous Driving

📅 2025-06-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current autonomous driving synthetic datasets suffer from limited scale, low photorealism, single-modality representation, and poor task adaptability. To address these limitations, we introduce the first full-stack, high-fidelity, multimodal (RGB + LiDAR) synthetic dataset specifically designed for autonomous driving. It provides fine-grained annotations across 64 object classes and supports multiple downstream tasks—including 2D object detection, bird’s-eye-view (BEV) semantic segmentation, and LiDAR point-wise semantic segmentation. Our methodology integrates a high-precision urban simulation engine, a physically based rendering pipeline, NeRF-enhanced texture synthesis for improved realism, and a programmable LiDAR simulator—ensuring spatiotemporal synchronization across sensors and geometric-semantic consistency. Extensive evaluations demonstrate state-of-the-art performance on multiple benchmarks, surpassing existing synthetic baselines. The dataset is publicly released and has been widely adopted by the research community.

Technology Category

Application Category

📝 Abstract
As perception models continue to develop, the need for large-scale datasets increases. However, data annotation remains far too expensive to effectively scale and meet the demand. Synthetic datasets provide a solution to boost model performance with substantially reduced costs. However, current synthetic datasets remain limited in their scope, realism, and are designed for specific tasks and applications. In this work, we present RealDriveSim, a realistic multi-modal synthetic dataset for autonomous driving that not only supports popular 2D computer vision applications but also their LiDAR counterparts, providing fine-grained annotations for up to 64 classes. We extensively evaluate our dataset for a wide range of applications and domains, demonstrating state-of-the-art results compared to existing synthetic benchmarks. The dataset is publicly available at https://realdrivesim.github.io/.
Problem

Research questions and friction points this paper is trying to address.

High cost of large-scale annotated datasets for autonomous driving
Limited scope and realism in current synthetic datasets
Need for multi-modal datasets supporting 2D and LiDAR tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-modal synthetic dataset for autonomous driving
Supports 2D vision and LiDAR applications
Provides fine-grained annotations for 64 classes
🔎 Similar Papers
No similar papers found.
A
Arpit Jadon
German Aerospace Center, Braunschweig, Germany
H
Haoran Wang
Max Planck Institute for Informatics, Saarland Informatics Campus, Germany
P
Phillip Thomas
Parallel Domain, San Francisco, CA, USA
Michael Stanley
Michael Stanley
Analytical Mechanics Associates
StatisticsProbabilityOptimizationInverse Problems
S
S. Nathaniel Cibik
Parallel Domain, San Francisco, CA, USA
R
Rachel Laurat
Parallel Domain, San Francisco, CA, USA
O
Omar Maher
Parallel Domain, San Francisco, CA, USA
Lukas Hoyer
Lukas Hoyer
ETH Zürich
Domain AdaptationDomain GeneralizationSemi-Supervised LearningSemantic Segmentation
Ozan Unal
Ozan Unal
ETH Zurich, Huawei Technologies
3D Semantic Segmentation3D Visual Grounding
Dengxin Dai
Dengxin Dai
Director of Research, Huawei Zurich Research Center
Computer VisionAutonomous Driving