Reusing Attention for One-stage Lane Topology Understanding

📅 2025-07-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address low accuracy and slow inference in lane topology understanding for autonomous driving, this paper proposes a single-stage end-to-end framework that jointly predicts traffic elements, lane centerlines, and their topological relationships. The core innovation lies in reusing the attention mechanisms within the Transformer decoder to explicitly model topological dependencies—eliminating error propagation and redundant computation inherent in two-stage approaches. It is the first work to enable cross-task sharing of attention resources from the detection module within a single-stage architecture. Additionally, we introduce a graph-free knowledge distillation strategy that transfers prior knowledge from standard HD maps to a lightweight, map-agnostic model. Evaluated on OpenLane-V2, our method achieves new state-of-the-art performance, improving AP by +3.2% while accelerating inference by 2.1×. The source code is publicly available.

Technology Category

Application Category

📝 Abstract
Understanding lane toplogy relationships accurately is critical for safe autonomous driving. However, existing two-stage methods suffer from inefficiencies due to error propagations and increased computational overheads. To address these challenges, we propose a one-stage architecture that simultaneously predicts traffic elements, lane centerlines and topology relationship, improving both the accuracy and inference speed of lane topology understanding for autonomous driving. Our key innovation lies in reusing intermediate attention resources within distinct transformer decoders. This approach effectively leverages the inherent relational knowledge within the element detection module to enable the modeling of topology relationships among traffic elements and lanes without requiring additional computationally expensive graph networks. Furthermore, we are the first to demonstrate that knowledge can be distilled from models that utilize standard definition (SD) maps to those operates without using SD maps, enabling superior performance even in the absence of SD maps. Extensive experiments on the OpenLane-V2 dataset show that our approach outperforms baseline methods in both accuracy and efficiency, achieving superior results in lane detection, traffic element identification, and topology reasoning. Our code is available at https://github.com/Yang-Li-2000/one-stage.git.
Problem

Research questions and friction points this paper is trying to address.

Improving accuracy and speed of lane topology understanding
Reusing attention to avoid expensive graph networks
Enabling performance without standard definition maps
Innovation

Methods, ideas, or system contributions that make the work stand out.

One-stage architecture predicts elements and topology simultaneously
Reuses attention in transformer decoders for efficiency
Distills knowledge from SD maps to non-SD models
🔎 Similar Papers
No similar papers found.
Y
Yang Li
Institute for AI Industry Research, Tsinghua University
Z
Zongzheng Zhang
Institute for AI Industry Research, Tsinghua University
X
Xuchong Qiu
Bosch Corporate Research, China
X
Xinrun Li
Bosch Corporate Research, China
Z
Ziming Liu
Bosch Corporate Research, China
L
Leichen Wang
Bosch Corporate Research, China
R
Ruikai Li
State Key Lab of Intelligent Transportation System, Beihang University
Zhenxin Zhu
Zhenxin Zhu
Xiaomi AD
AIGCNeRF
Huan-ang Gao
Huan-ang Gao
Ph.D. student, Tsinghua University
AgentVision & Robotics
X
Xiaojian Lin
Institute for AI Industry Research, Tsinghua University
Zhiyong Cui
Zhiyong Cui
Professor, Beihang University
Foundation ModelsAutonomous DrivingUrban ComputingTraffic PredictionTraffic Control
H
Hang Zhao
Institute for Interdisciplinary Information Sciences, Tsinghua University
H
Hao Zhao
Institute for AI Industry Research, Tsinghua University