π€ AI Summary
Existing autonomous driving world models primarily model only ego-vehicle trajectories, neglecting surrounding vehiclesβ motion, thereby compromising scene interaction fidelity; moreover, multi-vehicle trajectory matching and controllable generation in video remain challenging. This paper proposes EOT-WM, a unified driving world model that, for the first time, jointly models ego- and other-vehicle trajectories in the video latent space. It achieves high-fidelity, controllable video generation via BEV-to-image coordinate alignment, a spatiotemporal variational autoencoder, and a trajectory-injected diffusion Transformer. Innovatively, it maps multi-vehicle trajectories uniformly into the latent space and introduces a novel controllability metric based on similarity of control latent variables. Evaluated on nuScenes, EOT-WM reduces FID by 30% and FVD by 55% over prior methods, and enables unknown-scenario prediction conditioned on self-generated trajectories.
π Abstract
Advanced end-to-end autonomous driving systems predict other vehicles' motions and plan ego vehicle's trajectory. The world model that can foresee the outcome of the trajectory has been used to evaluate the end-to-end autonomous driving system. However, existing world models predominantly emphasize the trajectory of the ego vehicle and leave other vehicles uncontrollable. This limitation hinders their ability to realistically simulate the interaction between the ego vehicle and the driving scenario. In addition, it remains a challenge to match multiple trajectories with each vehicle in the video to control the video generation. To address above issues, a driving extbf{W}orld extbf{M}odel named EOT-WM is proposed in this paper, unifying extbf{E}go- extbf{O}ther vehicle extbf{T}rajectories in videos. Specifically, we first project ego and other vehicle trajectories in the BEV space into the image coordinate to match each trajectory with its corresponding vehicle in the video. Then, trajectory videos are encoded by the Spatial-Temporal Variational Auto Encoder to align with driving video latents spatially and temporally in the unified visual space. A trajectory-injected diffusion Transformer is further designed to denoise the noisy video latents for video generation with the guidance of ego-other vehicle trajectories. In addition, we propose a metric based on control latent similarity to evaluate the controllability of trajectories. Extensive experiments are conducted on the nuScenes dataset, and the proposed model outperforms the state-of-the-art method by 30% in FID and 55% in FVD. The model can also predict unseen driving scenes with self-produced trajectories.