Rolling Ahead Diffusion for Traffic Scene Simulation

๐Ÿ“… 2025-02-13
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing traffic scene simulation methods struggle to balance responsiveness and computational efficiency: diffusion models offer flexibility but incur high per-frame regeneration costs, while autoregressive models are computationally efficient yet lack long-horizon planning capability. This paper proposes Rolling Diffusion, a novel rolling-horizon diffusion generative framework featuring an innovative โ€œahead diffusionโ€ mechanism. In a single inference step, it jointly outputs the next-timestep deterministic trajectory and multiple subsequent noisy future states, thereby unifying Model Predictive Control (MPC)-style real-time responsiveness with autoregressive-style inference efficiency. Leveraging rolling time-window modeling, partial noise masking, and joint multi-agent motion modeling, Rolling Diffusion achieves 98.7% trajectory plausibility and high-fidelity real-time interaction on nuScenes simulation. It accelerates full-sequence diffusion MPC by 5.3ร— and significantly outperforms pure autoregressive baselines in long-horizon planning quality.

Technology Category

Application Category

๐Ÿ“ Abstract
Realistic driving simulation requires that NPCs not only mimic natural driving behaviors but also react to the behavior of other simulated agents. Recent developments in diffusion-based scenario generation focus on creating diverse and realistic traffic scenarios by jointly modelling the motion of all the agents in the scene. However, these traffic scenarios do not react when the motion of agents deviates from their modelled trajectories. For example, the ego-agent can be controlled by a stand along motion planner. To produce reactive scenarios with joint scenario models, the model must regenerate the scenario at each timestep based on new observations in a Model Predictive Control (MPC) fashion. Although reactive, this method is time-consuming, as one complete possible future for all NPCs is generated per simulation step. Alternatively, one can utilize an autoregressive model (AR) to predict only the immediate next-step future for all NPCs. Although faster, this method lacks the capability for advanced planning. We present a rolling diffusion based traffic scene generation model which mixes the benefits of both methods by predicting the next step future and simultaneously predicting partially noised further future steps at the same time. We show that such model is efficient compared to diffusion model based AR, achieving a beneficial compromise between reactivity and computational efficiency.
Problem

Research questions and friction points this paper is trying to address.

Simulate reactive NPC driving behaviors
Balance reactivity and computational efficiency
Predict immediate and future NPC movements
Innovation

Methods, ideas, or system contributions that make the work stand out.

Rolling diffusion model
Combines AR and MPC
Efficient reactive simulation
๐Ÿ”Ž Similar Papers