An End-to-End Learning-Based Multi-Sensor Fusion for Autonomous Vehicle Localization

📅 2025-03-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limitations of conventional multi-sensor fusion for autonomous vehicle localization—namely, its reliance on Gaussian assumptions, manual hyperparameter tuning, and poor generalization to long-tail scenarios—this paper proposes an end-to-end learnable fusion framework that jointly regresses ego-vehicle pose directly from raw sensor inputs, eliminating explicit uncertainty modeling and hand-crafted priors. Our key contributions are: (1) a high-order feature-based sensor encoder; (2) a customized end-to-end fusion network; and (3) a jointly optimized pose regression architecture that implicitly encodes feature-level uncertainty. Evaluated on real-world road scenes, our method achieves an 18.7% improvement in localization accuracy and demonstrates significantly enhanced robustness compared to both classical filters (EKF, UKF) and state-of-the-art learning-based approaches. Code and experimental validation video are available at https://youtu.be/q4iuobMbjME.

Technology Category

Application Category

📝 Abstract
Multi-sensor fusion is essential for autonomous vehicle localization, as it is capable of integrating data from various sources for enhanced accuracy and reliability. The accuracy of the integrated location and orientation depends on the precision of the uncertainty modeling. Traditional methods of uncertainty modeling typically assume a Gaussian distribution and involve manual heuristic parameter tuning. However, these methods struggle to scale effectively and address long-tail scenarios. To address these challenges, we propose a learning-based method that encodes sensor information using higher-order neural network features, thereby eliminating the need for uncertainty estimation. This method significantly eliminates the need for parameter fine-tuning by developing an end-to-end neural network that is specifically designed for multi-sensor fusion. In our experiments, we demonstrate the effectiveness of our approach in real-world autonomous driving scenarios. Results show that the proposed method outperforms existing multi-sensor fusion methods in terms of both accuracy and robustness. A video of the results can be viewed at https://youtu.be/q4iuobMbjME.
Problem

Research questions and friction points this paper is trying to address.

Improves autonomous vehicle localization accuracy
Eliminates manual parameter tuning in sensor fusion
Addresses long-tail scenarios in uncertainty modeling
Innovation

Methods, ideas, or system contributions that make the work stand out.

End-to-end neural network for sensor fusion
Higher-order neural network encodes sensor data
Eliminates manual parameter tuning and uncertainty estimation
🔎 Similar Papers
No similar papers found.
C
Changhong Lin
DiDi Autonomous Driving, DiDi Chuxing, Beijing, China
Jiarong Lin
Jiarong Lin
Associate Professor, Beihang University
RoboticsLiDAR SLAMSensor fusion3D reconstruction
Zhiqiang Sui
Zhiqiang Sui
Nuro
Perception for ManipulationProbabilistic RoboticsState Estimation
X
XiaoZhi Qu
DiDi Autonomous Driving, DiDi Chuxing, Beijing, China
R
Rui Wang
DiDi Autonomous Driving, DiDi Chuxing, Beijing, China
K
Kehua Sheng
DiDi Autonomous Driving, DiDi Chuxing, Beijing, China
B
Bo Zhang
DiDi Autonomous Driving, DiDi Chuxing, Beijing, China