GaussianCaR: Gaussian Splatting for Efficient Camera-Radar Fusion

📅 2026-02-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges of large viewpoint discrepancies and inefficient fusion between visual and radar modalities in autonomous driving by proposing GaussianCaR, an end-to-end BEV (bird’s-eye-view) semantic segmentation network. GaussianCaR introduces Gaussian splatting as a universal view transformer for the first time, enabling unified mapping of image pixels and radar points into BEV space without relying on complex geometric assumptions. The method further integrates multi-scale feature fusion with a Transformer decoder to efficiently extract rich BEV representations. Evaluated on the nuScenes dataset, GaussianCaR achieves IoU scores of 57.3%, 82.9%, and 50.1% for vehicles, roads, and lane markings, respectively—matching or surpassing state-of-the-art performance while delivering a 3.2× speedup in inference time.

Technology Category

Application Category

📝 Abstract
Robust and accurate perception of dynamic objects and map elements is crucial for autonomous vehicles performing safe navigation in complex traffic scenarios. While vision-only methods have become the de facto standard due to their technical advances, they can benefit from effective and cost-efficient fusion with radar measurements. In this work, we advance fusion methods by repurposing Gaussian Splatting as an efficient universal view transformer that bridges the view disparity gap, mapping both image pixels and radar points into a common Bird's-Eye View (BEV) representation. Our main contribution is GaussianCaR, an end-to-end network for BEV segmentation that, unlike prior BEV fusion methods, leverages Gaussian Splatting to map raw sensor information into latent features for efficient camera-radar fusion. Our architecture combines multi-scale fusion with a transformer decoder to efficiently extract BEV features. Experimental results demonstrate that our approach achieves performance on par with, or even surpassing, the state of the art on BEV segmentation tasks (57.3%, 82.9%, and 50.1% IoU for vehicles, roads, and lane dividers) on the nuScenes dataset, while maintaining a 3.2x faster inference runtime. Code and project page are available online.
Problem

Research questions and friction points this paper is trying to address.

camera-radar fusion
Bird's-Eye View
autonomous driving
BEV segmentation
sensor fusion
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gaussian Splatting
Camera-Radar Fusion
Bird's-Eye View (BEV)
View Transformation
Autonomous Driving Perception
🔎 Similar Papers
No similar papers found.
S
Santiago Montiel-Marín
Department of Electronics, University of Alcalá, Spain
M
Miguel Antunes-García
Department of Electronics, University of Alcalá, Spain
F
Fabio Sánchez-García
Department of Electronics, University of Alcalá, Spain
A
Angel Llamazares
Department of Electronics, University of Alcalá, Spain
Holger Caesar
Holger Caesar
Assistant Professor at TU Delft
Autonomous VehiclesComputer Vision
L
Luis M. Bergasa
Department of Electronics, University of Alcalá, Spain