Seeing the Unseen in Low-light Spike Streams

📅 2025-09-27
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address severe noise and extreme signal sparsity in event-based high-speed imaging under low-light conditions, this paper proposes Diff-SPK—the first work to introduce diffusion models into spike-based camera reconstruction. Methodologically, we design an Event-Time Feature Integration (ETFI) module that explicitly encodes inter-spike temporal intervals and aggregates sparse cross-frame signals; we further incorporate a ControlNet architecture to enable generative prior-guided conditional reconstruction. Our contributions are threefold: (1) a novel diffusion-based generative prior modeling paradigm for spike stream reconstruction; (2) the first real-world low-light spike stream benchmark dataset; and (3) state-of-the-art performance—demonstrating significant quantitative and qualitative improvements over existing methods, with reconstructed images exhibiting enhanced clarity, richer texture preservation, and more accurate motion detail recovery.

Technology Category

Application Category

📝 Abstract
Spike camera, a type of neuromorphic sensor with high-temporal resolution, shows great promise for high-speed visual tasks. Unlike traditional cameras, spike camera continuously accumulates photons and fires asynchronous spike streams. Due to unique data modality, spike streams require reconstruction methods to become perceptible to the human eye. However, lots of methods struggle to handle spike streams in low-light high-speed scenarios due to severe noise and sparse information. In this work, we propose Diff-SPK, the first diffusion-based reconstruction method for spike camera. Diff-SPK effectively leverages generative priors to supplement texture information in low-light conditions. Specifically, it first employs an extbf{E}nhanced extbf{T}exture extbf{f}rom Inter-spike extbf{I}nterval (ETFI) to aggregate sparse information from low-light spike streams. Then, ETFI serves as a conditioning input for ControlNet to generate the high-speed scenes. To improve the quality of results, we introduce an ETFI-based feature fusion module during the generation process. Moreover, we establish the first bona fide benchmark for the low-light spike stream reconstruction task. It significantly surpasses existing reconstruction datasets in scale and provides quantitative illumination information. The performance on real low-light spike streams demonstrates the superiority of Diff-SPK.
Problem

Research questions and friction points this paper is trying to address.

Reconstructing low-light spike streams with severe noise
Handling sparse information in high-speed visual scenarios
Generating perceptible images from asynchronous spike data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses diffusion model for spike stream reconstruction
Employs ETFI to aggregate sparse spike information
Integrates ControlNet with ETFI feature fusion module
🔎 Similar Papers
No similar papers found.
L
Liwen Hu
School of Computer Science, Peking University
Y
Yang Li
School of Computer Science, Peking University
M
Mianzhi Liu
National Biomedical Imaging Center, Peking University
Yijia Guo
Yijia Guo
Peking University
3DV
Shenghao Xie
Shenghao Xie
Ph.D. Student, AAIS, PKU
Computer VisionMachine Learning
Ziluo Ding
Ziluo Ding
Unknown affiliation
Reinforcement LearningOptical Flow
Tiejun Huang
Tiejun Huang
Professor,School of Computer Science, Peking University
Visual Information Processing
L
Lei Ma
National Biomedical Imaging Center, Peking University