DST-Calib: A Dual-Path, Self-Supervised, Target-Free LiDAR-Camera Extrinsic Calibration Network

📅 2026-01-03
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of LiDAR–camera extrinsic calibration, which typically relies on manual calibration targets or specific static scenes and is thus difficult to deploy online. To overcome this limitation, we propose the first self-supervised extrinsic calibration network that operates without any calibration targets. Our method employs a dual-path architecture with differential feature maps, replacing conventional two-branch designs to enhance cross-modal feature association while reducing model complexity. Furthermore, we introduce a depth map–based multi-view camera augmentation strategy to improve generalization and enable online adaptive calibration. Extensive experiments on five public benchmarks and our own collected dataset demonstrate that the proposed approach significantly outperforms existing methods, achieving state-of-the-art performance in both calibration accuracy and generalization capability.

Technology Category

Application Category

📝 Abstract
LiDAR-camera extrinsic calibration is essential for multi-modal data fusion in robotic perception systems. However, existing approaches typically rely on handcrafted calibration targets (e.g., checkerboards) or specific, static scene types, limiting their adaptability and deployment in real-world autonomous and robotic applications. This article presents the first self-supervised LiDAR-camera extrinsic calibration network that operates in an online fashion and eliminates the need for specific calibration targets. We first identify a significant generalization degradation problem in prior methods, caused by the conventional single-sided data augmentation strategy. To overcome this limitation, we propose a novel double-sided data augmentation technique that generates multi-perspective camera views using estimated depth maps, thereby enhancing robustness and diversity during training. Built upon this augmentation strategy, we design a dual-path, self-supervised calibration framework that reduces the dependence on high-precision ground truth labels and supports fully adaptive online calibration. Furthermore, to improve cross-modal feature association, we replace the traditional dual-branch feature extraction design with a difference map construction process that explicitly correlates LiDAR and camera features. This not only enhances calibration accuracy but also reduces model complexity. Extensive experiments conducted on five public benchmark datasets, as well as our own recorded dataset, demonstrate that the proposed method significantly outperforms existing approaches in terms of generalizability.
Problem

Research questions and friction points this paper is trying to address.

LiDAR-camera extrinsic calibration
calibration targets
self-supervised
multi-modal data fusion
generalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

self-supervised calibration
target-free
dual-path network
double-sided data augmentation
cross-modal feature association
🔎 Similar Papers
No similar papers found.
Z
Zhiwei Huang
Department of Control Science & Engineering, the College of Electronics & Information Engineering, Tongji University, Shanghai 201804, China
Yanwei Fu
Yanwei Fu
Fudan University
Computer visionmachine learningMultimedia
Yi Zhou
Yi Zhou
Professor at Hunan University; Director of Neuromorphic Automation and Intelligence Lab (NAIL)
VO/SLAMEvent-based VisionHigh-Performance Neuromorphic ComputationMulti-view geometry
Xieyuanli Chen
Xieyuanli Chen
Associate Professor, NUDT, China
RoboticsSLAMLocalizationLiDAR PerceptionRobot Learning
Q
Qijun Chen
R
Rui Fan
Department of Control Science & Engineering, the College of Electronics & Information Engineering, Shanghai Research Institute for Intelligent Autonomous Systems, the State Key Laboratory of Intelligent Autonomous Systems, and Frontiers Science Center for Intelligent Autonomous Systems, Tongji University, Shanghai 201804, China, as well as with the National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Xi’an 71004