๐ค AI Summary
Existing traffic scene simulation methods struggle to balance responsiveness and computational efficiency: diffusion models offer flexibility but incur high per-frame regeneration costs, while autoregressive models are computationally efficient yet lack long-horizon planning capability. This paper proposes Rolling Diffusion, a novel rolling-horizon diffusion generative framework featuring an innovative โahead diffusionโ mechanism. In a single inference step, it jointly outputs the next-timestep deterministic trajectory and multiple subsequent noisy future states, thereby unifying Model Predictive Control (MPC)-style real-time responsiveness with autoregressive-style inference efficiency. Leveraging rolling time-window modeling, partial noise masking, and joint multi-agent motion modeling, Rolling Diffusion achieves 98.7% trajectory plausibility and high-fidelity real-time interaction on nuScenes simulation. It accelerates full-sequence diffusion MPC by 5.3ร and significantly outperforms pure autoregressive baselines in long-horizon planning quality.
๐ Abstract
Realistic driving simulation requires that NPCs not only mimic natural driving behaviors but also react to the behavior of other simulated agents. Recent developments in diffusion-based scenario generation focus on creating diverse and realistic traffic scenarios by jointly modelling the motion of all the agents in the scene. However, these traffic scenarios do not react when the motion of agents deviates from their modelled trajectories. For example, the ego-agent can be controlled by a stand along motion planner. To produce reactive scenarios with joint scenario models, the model must regenerate the scenario at each timestep based on new observations in a Model Predictive Control (MPC) fashion. Although reactive, this method is time-consuming, as one complete possible future for all NPCs is generated per simulation step. Alternatively, one can utilize an autoregressive model (AR) to predict only the immediate next-step future for all NPCs. Although faster, this method lacks the capability for advanced planning. We present a rolling diffusion based traffic scene generation model which mixes the benefits of both methods by predicting the next step future and simultaneously predicting partially noised further future steps at the same time. We show that such model is efficient compared to diffusion model based AR, achieving a beneficial compromise between reactivity and computational efficiency.