🤖 AI Summary
Motion blur severely degrades image quality, hindering real-time edge applications such as autonomous driving and UAV perception. To address this, we propose a lightweight U-shaped deblurring network tailored for resource-constrained edge devices. Our method introduces three key innovations: (1) a Lightweight Deblurring (LD) block to reduce computational redundancy; (2) a Multi-Level Integrated Aggregation (MLIA) module to enhance multi-scale feature representation; and (3) a Cross-source Fusion (X-Fuse) block enabling efficient encoder–decoder information interaction. The resulting model achieves a compact footprint of only 5.85 million parameters and 15.76 GMACs, with single-frame inference latency of 6 ms (<7 ms) on both GPU and mobile platforms—exceeding 140 FPS. It attains a PSNR of 30.67 dB, striking a strong balance between accuracy and speed. The source code is publicly available.
📝 Abstract
Motion blur caused by camera or object movement severely degrades image quality and poses challenges for real-time applications such as autonomous driving, UAV perception, and medical imaging. In this paper, a lightweight U-shaped network tailored for real-time deblurring is presented and named RT-Focuser. To balance speed and accuracy, we design three key components: Lightweight Deblurring Block (LD) for edge-aware feature extraction, Multi-Level Integrated Aggregation module (MLIA) for encoder integration, and Cross-source Fusion Block (X-Fuse) for progressive decoder refinement. Trained on a single blurred input, RT-Focuser achieves 30.67 dB PSNR with only 5.85M parameters and 15.76 GMACs. It runs 6ms per frame on GPU and mobile, exceeds 140 FPS on both, showing strong potential for deployment on the edge. The official code and usage are available on: https://github.com/ReaganWu/RT-Focuser.