🤖 AI Summary
This work addresses unified image restoration across multiple degradation types (e.g., denoising, deraining, dehazing) with a novel frequency-domain modeling framework—the first single-model, all-in-one solution. Methodologically, it introduces a dual-Transformer architecture: a frequency-aware Degradation Estimator (Dformer) and a degradation-adaptive Restorer (Rformer), operating in synergy. Crucially, it explicitly decouples degradation modeling into the frequency domain (via DCT or wavelet transforms) and leverages this decomposition to condition self-attention mechanisms for selective band-wise restoration. This enables joint learning of degradation representations and frequency-band-adaptive reconstruction. The proposed method achieves state-of-the-art performance across five major restoration tasks. Moreover, it demonstrates significantly improved generalization to real-world degradations, spatially variant degradations, and unseen degradation intensities—marking a substantial advance in unified, frequency-guided image restoration.
📝 Abstract
This work aims to tackle the all-in-one image restoration task, which seeks to handle multiple types of degradation with a single model. The primary challenge is to extract degradation representations from the input degraded images and use them to guide the model's adaptation to specific degradation types. Building on the insight that various degradations affect image content differently across frequency bands, we propose a new dual-transformer approach comprising two components: a frequency-aware Degradation estimation transformer (Dformer) and a degradation-adaptive Restoration transformer (Rformer). The Dformer captures the essential characteristics of various degradations by decomposing the input into different frequency components. By understanding how degradations affect these frequency components, the Dformer learns robust priors that effectively guide the restoration process. The Rformer then employs a degradation-adaptive self-attention module to selectively focus on the most affected frequency components, guided by the learned degradation representations. Extensive experimental results demonstrate that our approach outperforms existing methods in five representative restoration tasks, including denoising, deraining, dehazing, deblurring, and low-light enhancement. Additionally, our method offers benefits for handling, real-world degradations, spatially variant degradations, and unseen degradation levels.