🤖 AI Summary
To address the challenge of modeling human driving behavior diversity in autonomous driving simulation, this paper proposes a discretized contrastive learning framework. First, it integrates contrastive learning with vector quantization (VQ) to learn an interpretable, quantifiable driving style dictionary directly from real-world driving data. Second, it constructs a style-conditioned diffusion policy model to generate high-fidelity, safe, and human-like driving behaviors. The method unifies behavioral cloning, VQ-based representation learning, and conditional diffusion modeling, enabling controllable and interpretable behavior synthesis. Experiments demonstrate that the generated behaviors reduce collision rates by 23% and improve expert-rated human-likeness scores by 31% in simulation, significantly outperforming mainstream machine learning baselines. This advancement enhances both the realism of autonomous driving testing and the validity of evaluation metrics.
📝 Abstract
Learning to perform accurate and rich simulations of human driving behaviors from data for autonomous vehicle testing remains challenging due to human driving styles' high diversity and variance. We address this challenge by proposing a novel approach that leverages contrastive learning to extract a dictionary of driving styles from pre-existing human driving data. We discretize these styles with quantization, and the styles are used to learn a conditional diffusion policy for simulating human drivers. Our empirical evaluation confirms that the behaviors generated by our approach are both safer and more human-like than those of the machine-learning-based baseline methods. We believe this has the potential to enable higher realism and more effective techniques for evaluating and improving the performance of autonomous vehicles.