What Matters for Scalable and Robust Learning in End-to-End Driving Planners?

📅 2026-03-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge that existing end-to-end autonomous driving methods, while performing well on open-loop datasets, struggle to balance scalability and robustness in closed-loop settings. The authors systematically evaluate the impact of key design choices—including high-resolution perceptual representations, decoupled trajectory modeling, and generative planning—on closed-loop performance. They propose BevAD, a lightweight end-to-end framework based purely on behavioral cloning, which integrates bird’s-eye-view feature grids, high-fidelity perception, and a decoupled trajectory generation mechanism. Evaluated on the Bench2Drive benchmark, BevAD achieves a 72.7% success rate and demonstrates strong data-scaling capabilities, thereby revealing both the limitations of prevailing architectures in closed-loop scenarios and the potential for synergistic optimization across perception and planning components.

Technology Category

Application Category

📝 Abstract
End-to-end autonomous driving has gained significant attention for its potential to learn robust behavior in interactive scenarios and scale with data. Popular architectures often build on separate modules for perception and planning connected through latent representations, such as bird's eye view feature grids, to maintain end-to-end differentiability. This paradigm emerged mostly on open-loop datasets, with evaluation focusing not only on driving performance, but also intermediate perception tasks. Unfortunately, architectural advances that excel in open-loop often fail to translate to scalable learning of robust closed-loop driving. In this paper, we systematically re-examine the impact of common architectural patterns on closed-loop performance: (1) high-resolution perceptual representations, (2) disentangled trajectory representations, and (3) generative planning. Crucially, our analysis evaluates the combined impact of these patterns, revealing both unexpected limitations as well as underexplored synergies. Building on these insights, we introduce BevAD, a novel lightweight and highly scalable end-to-end driving architecture. BevAD achieves 72.7% success rate on the Bench2Drive benchmark and demonstrates strong data-scaling behavior using pure imitation learning. Our code and models are publicly available here: https://dmholtz.github.io/bevad/
Problem

Research questions and friction points this paper is trying to address.

end-to-end driving
closed-loop performance
scalable learning
robustness
autonomous driving
Innovation

Methods, ideas, or system contributions that make the work stand out.

end-to-end driving
closed-loop performance
BEV representation
scalable imitation learning
generative planning
🔎 Similar Papers
No similar papers found.