DPGP: A Hybrid 2D-3D Dual Path Potential Ghost Probe Zone Prediction Framework for Safe Autonomous Driving

📅 2025-04-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In dense urban traffic, “ghost detection”—i.e., sudden emergence of pedestrians or objects into the vehicle’s path from non-line-of-sight (NLOS) regions—poses critical safety risks. Method: This paper proposes a lightweight 2D–3D fusion prediction framework leveraging only a monocular camera. It extends ghost detection region prediction to non-vehicular dynamic objects for the first time and introduces an unsupervised multi-representation feature fusion mechanism grounded in depth discontinuity priors—requiring no V2X communication, specialized hardware, or manual depth annotations. Contribution/Results: Evaluated on a newly curated dataset of 12K annotated images, our model outperforms state-of-the-art methods with <1.2M parameters and real-time inference speed exceeding 30 FPS, demonstrating readiness for onboard deployment. Both code and dataset will be publicly released.

Technology Category

Application Category

📝 Abstract
Modern robots must coexist with humans in dense urban environments. A key challenge is the ghost probe problem, where pedestrians or objects unexpectedly rush into traffic paths. This issue affects both autonomous vehicles and human drivers. Existing works propose vehicle-to-everything (V2X) strategies and non-line-of-sight (NLOS) imaging for ghost probe zone detection. However, most require high computational power or specialized hardware, limiting real-world feasibility. Additionally, many methods do not explicitly address this issue. To tackle this, we propose DPGP, a hybrid 2D-3D fusion framework for ghost probe zone prediction using only a monocular camera during training and inference. With unsupervised depth prediction, we observe ghost probe zones align with depth discontinuities, but different depth representations offer varying robustness. To exploit this, we fuse multiple feature embeddings to improve prediction. To validate our approach, we created a 12K-image dataset annotated with ghost probe zones, carefully sourced and cross-checked for accuracy. Experimental results show our framework outperforms existing methods while remaining cost-effective. To our knowledge, this is the first work extending ghost probe zone prediction beyond vehicles, addressing diverse non-vehicle objects. We will open-source our code and dataset for community benefit.
Problem

Research questions and friction points this paper is trying to address.

Predict ghost probe zones for safe autonomous driving
Address ghost probe problem with hybrid 2D-3D fusion
Detect unexpected pedestrian or object intrusions in traffic
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hybrid 2D-3D fusion for ghost probe prediction
Unsupervised depth prediction with monocular camera
Multi-feature fusion enhances prediction robustness
🔎 Similar Papers
No similar papers found.
W
Weiming Qu
National Key Laboratory of General Artificial Intelligence, Key Laboratory of Machine Perception (MoE), School of Intelligence Science and Technology, Peking University, Beijing 100871, China; PKU-WUHAN Institute for Artificial Intelligence, Wuhan, China
Jiawei Du
Jiawei Du
National Taiwan University; ex-Intern @ Samsung Research
Speech processingNeural codingGenerative AIAI security
S
Shenghai Yuan
School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 639798, Singapore
J
Jia Wang
National Key Laboratory of General Artificial Intelligence, Key Laboratory of Machine Perception (MoE), School of Intelligence Science and Technology, Peking University, Beijing 100871, China
Y
Yang Sun
School of Instrument Science and Engineering, Southeast University, Nanjing 210018, China
S
Shengyi Liu
School of Instrument Science and Engineering, Southeast University, Nanjing 210018, China
Y
Yuanhao Zhu
China Automotive Innovation Corporation, Nanjing, China
Jianfeng Yu
Jianfeng Yu
China Automotive Innovation Corporation, Nanjing, China
Song Cao
Song Cao
University of Southern California
Computer Vision
R
Rui Xia
China Automotive Innovation Corporation, Nanjing, China
X
Xiaoyu Tang
School of Instrument Science and Engineering, Southeast University, Nanjing 210018, China
Xihong Wu
Xihong Wu
Peking University
Machine learningSpeech signal processingArtificial intelligence
D
Dingsheng Luo
National Key Laboratory of General Artificial Intelligence, Key Laboratory of Machine Perception (MoE), School of Intelligence Science and Technology, Peking University, Beijing 100871, China; PKU-WUHAN Institute for Artificial Intelligence, Wuhan, China