Toward Deep Representation Learning for Event-Enhanced Visual Autonomous Perception: the eAP Dataset

📅 2026-03-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the significant performance degradation of existing visual perception systems under challenging conditions such as low illumination and the absence of large-scale autonomous driving datasets that support the integration of event cameras with deep learning. To this end, the authors introduce eAP, the largest event-based autonomous driving dataset to date, and propose a geometry-aware representation learning framework that effectively incorporates event data into mainstream 3D vehicle detection networks for the first time. This integration substantially enhances robustness under complex lighting conditions. Furthermore, the proposed approach enables real-time, high-performance event-driven time-to-contact (TTC) estimation at up to 200 frames per second.

Technology Category

Application Category

📝 Abstract
Recent visual autonomous perception systems achieve remarkable performances with deep representation learning. However, they fail in scenarios with challenging illumination.While event cameras can mitigate this problem, there is a lack of a large-scale dataset to develop event-enhanced deep visual perception models in autonomous driving scenes. To address the gap, we present the eAP (event-enhanced Autonomous Perception) dataset, the largest dataset with event cameras for autonomous perception. We demonstrate how eAP can facilitate the study of different autonomous perception tasks, including 3D vehicle detection and object time-to-contact (TTC) estimation, through deep representation learning. Based on eAP, we demonstrate the ffrst successful use of events to improve a popular 3D vehicle detection network in challenging illumination scenarios. eAP also enables a devoted study of the representation learning problem of object TTC estimation. We show how a geometryaware representation learning framework leads to the best eventbased object TTC estimation network that operates at 200 FPS. The dataset, code, and pre-trained models will be made publicly available for future research.
Problem

Research questions and friction points this paper is trying to address.

event camera
autonomous perception
challenging illumination
large-scale dataset
deep representation learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

event camera
deep representation learning
autonomous perception
3D vehicle detection
time-to-contact estimation
🔎 Similar Papers
No similar papers found.
J
Jinghang Li
Neuromorphic Automation and Intelligence Lab (NAIL) at School of AI & Robotics, Hunan University, Changsha 410082, China
S
Shichao Li
ByteDance, Shenzhen 518063, China
Qing Lian
Qing Lian
HKUST
P
Peiliang Li
Zhuoyu Technology, Shenzhen 518055, China
Xiaozhi Chen
Xiaozhi Chen
ZYT
Machine LearningComputer Vision
Yi Zhou
Yi Zhou
Professor at Hunan University; Director of Neuromorphic Automation and Intelligence Lab (NAIL)
VO/SLAMEvent-based VisionHigh-Performance Neuromorphic ComputationMulti-view geometry