HazyDet: Open-source Benchmark for Drone-view Object Detection with Depth-cues in Hazy Scenes

📅 2024-09-30
🏛️ arXiv.org
📈 Citations: 4
Influential: 1
📄 PDF
🤖 AI Summary
To address the slow progress in object detection for foggy drone-captured imagery—largely due to the absence of a dedicated benchmark—this paper introduces HazyDet, the first open-source foggy drone detection benchmark comprising 383K real and synthetically fogged images. We further propose DeCoDet, a depth-aware detector designed specifically for this challenging domain. Its key contributions include: (1) a novel detection paradigm integrating depth priors; (2) a multi-scale depth-aware head coupled with a dynamic depth-conditioned convolution module; and (3) a scale-invariant reconstruction loss enabling pseudo-label–driven self-supervised learning of depth cues. Extensive experiments on HazyDet demonstrate that DeCoDet significantly outperforms state-of-the-art methods, especially for small and distant objects. All data, models, and code are publicly released to foster community advancement.

Technology Category

Application Category

📝 Abstract
Drone-based object detection in adverse weather conditions is crucial for enhancing drones' environmental perception, yet it remains largely unexplored due to the lack of relevant benchmarks. To bridge this gap, we introduce HazyDet, a large-scale dataset tailored for drone-based object detection in hazy scenes. It encompasses 383,000 real-world instances, collected from both naturally hazy environments and normal scenes with synthetically imposed haze effects to simulate adverse weather conditions. By observing the significant variations in object scale and clarity under different depth and haze conditions, we designed a Depth Conditioned Detector (DeCoDet) to incorporate this prior knowledge. DeCoDet features a Multi-scale Depth-aware Detection Head that seamlessly integrates depth perception, with the resulting depth cues harnessed by a dynamic Depth Condition Kernel module. Furthermore, we propose a Scale Invariant Refurbishment Loss to facilitate the learning of robust depth cues from pseudo-labels. Extensive evaluations on the HazyDet dataset demonstrate the flexibility and effectiveness of our method, yielding significant performance improvements. Our dataset and toolkit are available at https://github.com/GrokCV/HazyDet.
Problem

Research questions and friction points this paper is trying to address.

Object detection in hazy drone-view scenes lacks benchmarks
Haze causes severe visual degradation in aerial images
Existing methods struggle with synthetic-to-real domain shifts
Innovation

Methods, ideas, or system contributions that make the work stand out.

Depth-Conditioned Kernel for dynamic feature modulation
Progressive Domain Fine-Tuning for synthetic-to-real adaptation
Scale-Invariant Refurbishment Loss for noisy depth learning
🔎 Similar Papers
No similar papers found.
C
Changfeng Feng
Shijiazhuang Campus, Army Engineering University of PLA
Zhenyuan Chen
Zhenyuan Chen
Nankai University
Computer VisionVision-Language Models
R
Ren-Jay Kou
Shijiazhuang Campus, Army Engineering University of PLA
Guangwei Gao
Guangwei Gao
Professor of PCALab@NJUST, IEEE/CCF/CSIG/CAAI/CAA Senior Member
Pattern RecognitionImage UnderstandingMachine Learning
Chunping Wang
Chunping Wang
Finvolution
machine learning
X
Xiang Li
VCIP, CS, Nankai University
X
Xiangbo Shu
School of Computer Science and Engineering, Nanjing University of Science and Technology
Y
Yimian Dai
School of Computer Science and Engineering, Nanjing University of Science and Technology
Q
Qiang Fu
Shijiazhuang Campus, Army Engineering University of PLA
J
Jian Yang
School of Computer Science and Engineering, Nanjing University of Science and Technology