CompEvent: Complex-valued Event-RGB Fusion for Low-light Video Enhancement and Deblurring

📅 2025-11-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Low-light video deblurring faces dual degradation challenges—insufficient illumination and motion blur—particularly in nighttime surveillance and autonomous driving. Existing two-stage fusion approaches suffer from limited modeling capacity for jointly addressing these coupled degradations. To this end, this paper proposes the first end-to-end framework for fusing event-camera data with RGB video. Our core innovation lies in a complex-valued neural network-driven spatiotemporal fusion mechanism, integrating complex convolution, complex spatiotemporal alignment GRUs, and a complex-domain spatial-frequency joint learning module. This enables continuous, unified alignment and deep fusion of event streams and RGB frames within a single complex-valued representation space. Extensive experiments demonstrate state-of-the-art performance across multiple benchmarks, achieving average improvements of 1.82 dB in PSNR and 0.023 in SSIM over prior methods. The code is publicly released, ensuring strong reproducibility and practical applicability.

Technology Category

Application Category

📝 Abstract
Low-light video deblurring poses significant challenges in applications like nighttime surveillance and autonomous driving due to dim lighting and long exposures. While event cameras offer potential solutions with superior low-light sensitivity and high temporal resolution, existing fusion methods typically employ staged strategies, limiting their effectiveness against combined low-light and motion blur degradations. To overcome this, we propose CompEvent, a complex neural network framework enabling holistic full-process fusion of event data and RGB frames for enhanced joint restoration. CompEvent features two core components: 1) Complex Temporal Alignment GRU, which utilizes complex-valued convolutions and processes video and event streams iteratively via GRU to achieve temporal alignment and continuous fusion; and 2) Complex Space-Frequency Learning module, which performs unified complex-valued signal processing in both spatial and frequency domains, facilitating deep fusion through spatial structures and system-level characteristics. By leveraging the holistic representation capability of complex-valued neural networks, CompEvent achieves full-process spatiotemporal fusion, maximizes complementary learning between modalities, and significantly strengthens low-light video deblurring capability. Extensive experiments demonstrate that CompEvent outperforms SOTA methods in addressing this challenging task. The code is available at https://github.com/YuXie1/CompEvent.
Problem

Research questions and friction points this paper is trying to address.

Enhancing low-light video quality and removing motion blur
Fusing event camera data with RGB frames for joint restoration
Addressing combined degradations of dim lighting and blurring
Innovation

Methods, ideas, or system contributions that make the work stand out.

Complex neural network for event-RGB fusion
Complex Temporal Alignment GRU for continuous fusion
Complex Space-Frequency Learning in dual domains
🔎 Similar Papers
No similar papers found.
M
Mingchen Zhong
University of Science and Technology of China
X
Xin Lu
University of Science and Technology of China
D
Dong Li
University of Science and Technology of China
Senyan Xu
Senyan Xu
University of Science and Technology of China
Computer VisionImage Processing
R
Ruixuan Jiang
University of Science and Technology of China
X
Xueyang Fu
University of Science and Technology of China
Baocai Yin
Baocai Yin
Unknown affiliation