🤖 AI Summary
Existing modular autonomous driving systems decouple trajectory prediction from motion planning, hindering explicit modeling of inter-vehicle trajectory dependencies and limiting collaborative planning capabilities. To address this, we propose a microsimulation-based conditional trajectory prediction method: leveraging ego-vehicle-generated candidate trajectories as conditioning inputs, it jointly models surrounding agents’ responsive behaviors. Crucially, we introduce adversarial inverse reinforcement learning (AIRL)-trained behavior models into the simulation framework—enabling cross-agent conditional modeling and online adaptation of ego-vehicle trajectories. This approach breaks the conventional prediction-planning separation paradigm. Evaluated on complex interactive scenarios—including cooperative lane-changing and merging—the method significantly improves prediction fidelity, planning rationality, and safety while maintaining high interpretability.
📝 Abstract
Modular automated driving systems commonly handle prediction and planning as sequential, separate tasks, thereby prohibiting cooperative maneuvers. To enable cooperative planning, this work introduces a prediction model that models the conditional dependencies between trajectories. For this, predictions are generated by a microscopic traffic simulation, with the individual traffic participants being controlled by a realistic behavior model trained via Adversarial Inverse Reinforcement Learning. By assuming various candidate trajectories for the automated vehicle, we generate predictions conditioned on each of them. Furthermore, our approach allows the candidate trajectories to adapt dynamically during the prediction rollout. Several example scenarios are available at https://conditionalpredictionbysimulation.github.io/.