🤖 AI Summary
Unsupervised restoration of images degraded by multiple factors (e.g., blur, noise, rain streaks, compression artifacts) faces key bottlenecks: high computational cost from iterative search, reliance on labeled degradation types, and dependence on ground-truth reference images.
Method: We propose a lightweight sequential decision-making agent that introduces the first MLLM-driven no-reference perceptual reward mechanism—capable of assessing restoration quality without degradation labels or reference images. Our approach employs a policy gradient reinforcement learning framework, where the MLLM serves as a learnable evaluator and a lightweight network generates deterministic tool invocation sequences.
Contribution/Results: Under zero supervision, our method achieves state-of-the-art performance on full-reference metrics (e.g., PSNR, SSIM), surpasses all prior work on no-reference metrics (e.g., NIQE, BRISQUE), and significantly accelerates inference speed—demonstrating superior efficiency and generalizability for multi-degradation restoration.
📝 Abstract
Complex image restoration aims to recover high-quality images from inputs affected by multiple degradations such as blur, noise, rain, and compression artifacts. Recent restoration agents, powered by vision-language models and large language models, offer promising restoration capabilities but suffer from significant efficiency bottlenecks due to reflection, rollback, and iterative tool searching. Moreover, their performance heavily depends on degradation recognition models that require extensive annotations for training, limiting their applicability in label-free environments. To address these limitations, we propose a policy optimization-based restoration framework that learns an lightweight agent to determine tool-calling sequences. The agent operates in a sequential decision process, selecting the most appropriate restoration operation at each step to maximize final image quality. To enable training within label-free environments, we introduce a novel reward mechanism driven by multimodal large language models, which act as human-aligned evaluator and provide perceptual feedback for policy improvement. Once trained, our agent executes a deterministic restoration plans without redundant tool invocations, significantly accelerating inference while maintaining high restoration quality. Extensive experiments show that despite using no supervision, our method matches SOTA performance on full-reference metrics and surpasses existing approaches on no-reference metrics across diverse degradation scenarios.