SEED4D: A Synthetic Ego-Exo Dynamic 4D Data Generator, Driving Dataset and Benchmark

📅 2024-12-01
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing 3D/4D reconstruction methods for autonomous driving are severely limited by the scarcity of synchronized, multi-view, multi-temporal ego/exo RGB and LiDAR data in real-world settings. Method: This paper introduces the first synthetic dynamic multi-view 4D dataset generator and benchmark tailored for autonomous driving, featuring a customizable, open-source spatiotemporal multi-view synthesis framework. It supports physics-based sensor simulation, multi-camera calibration modeling, and configurations compatible with NuScenes, KITTI-360, and Waymo. Contribution/Results: We release 212K static images (2K scenes) and 16.8M dynamic images (10K trajectories × 100 frames), all with synchronized ego/exo RGB-LiDAR annotations. This benchmark fills a critical gap in multi-source temporal supervision for complex dynamic scenes, significantly improving generalization and cross-view consistency of few-shot 3D/4D reconstruction models, and advancing state-of-the-art performance across multiple benchmarks.

Technology Category

Application Category

📝 Abstract
Models for egocentric 3D and 4D reconstruction, including few-shot interpolation and extrapolation settings, can benefit from having images from exocentric viewpoints as supervision signals. No existing dataset provides the necessary mixture of complex, dynamic, and multi-view data. To facilitate the development of 3D and 4D reconstruction methods in the autonomous driving context, we propose a Synthetic Ego--Exo Dynamic 4D (SEED4D) data generator and dataset. We present a customizable, easy-to-use data generator for spatio-temporal multi-view data creation. Our open-source data generator allows the creation of synthetic data for camera setups commonly used in the NuScenes, KITTI360, and Waymo datasets. Additionally, SEED4D encompasses two large-scale multi-view synthetic urban scene datasets. Our static (3D) dataset encompasses 212k inward- and outward-facing vehicle images from 2k scenes, while our dynamic (4D) dataset contains 16.8M images from 10k trajectories, each sampled at 100 points in time with egocentric images, exocentric images, and LiDAR data. The datasets and the data generator can be found at https://seed4d.github.io/.
Problem

Research questions and friction points this paper is trying to address.

3D and 4D model reconstruction
autonomous driving technology
image acquisition from multiple angles and time points
Innovation

Methods, ideas, or system contributions that make the work stand out.

SEED4D
Multi-perspective Imaging
Autonomous Driving Data
🔎 Similar Papers
No similar papers found.