Degradation-Aware Image Enhancement via Vision-Language Classification

📅 2025-06-05
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
Real-world image degradations are diverse and challenging to identify automatically. Method: This paper pioneers the integration of vision-language models (VLMs) into degradation classification, proposing a fine-grained degradation-aware and modular restoration framework. It categorizes degradations into four types—super-resolution-related, reflection, motion blur, and no degradation—and employs a VLM to accurately classify each input image, subsequently activating a dedicated reconstruction network for on-demand enhancement. Domain adaptation is incorporated to improve cross-scenario generalization. Results: On multiple benchmark datasets, the method achieves 92.7% degradation classification accuracy and significantly outperforms unified enhancement approaches in PSNR and SSIM. It also enhances perceptual quality of restored images and improves performance on downstream tasks, thereby overcoming inherent limitations of end-to-end models in interpretability and generalizability.

Technology Category

Application Category

📝 Abstract
Image degradation is a prevalent issue in various real-world applications, affecting visual quality and downstream processing tasks. In this study, we propose a novel framework that employs a Vision-Language Model (VLM) to automatically classify degraded images into predefined categories. The VLM categorizes an input image into one of four degradation types: (A) super-resolution degradation (including noise, blur, and JPEG compression), (B) reflection artifacts, (C) motion blur, or (D) no visible degradation (high-quality image). Once classified, images assigned to categories A, B, or C undergo targeted restoration using dedicated models tailored for each specific degradation type. The final output is a restored image with improved visual quality. Experimental results demonstrate the effectiveness of our approach in accurately classifying image degradations and enhancing image quality through specialized restoration models. Our method presents a scalable and automated solution for real-world image enhancement tasks, leveraging the capabilities of VLMs in conjunction with state-of-the-art restoration techniques.
Problem

Research questions and friction points this paper is trying to address.

Classify image degradation types using Vision-Language Model
Restore degraded images with targeted models for each type
Improve visual quality via scalable automated enhancement framework
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Vision-Language Model for degradation classification
Applies dedicated models for specific degradation types
Automates scalable image enhancement via VLM
🔎 Similar Papers
No similar papers found.
J
Jie Cai
OPPO AI Center, Palo Alto, CA, US
Kangning Yang
Kangning Yang
InnoPeak Technology
Deep LearningMutimodal LearningAffective ComputingHuman-Computer Interaction
J
Jiaming Ding
OPPO AI Center, Palo Alto, CA, US
Lan Fu
Lan Fu
University of South Carolina
computer vision
O
Ouyang Ling
OPPO AI Center, Palo Alto, CA, US
J
Jiang Li
OPPO AI Center, Palo Alto, CA, US
J
Jinglin Shen
OPPO AI Center, Palo Alto, CA, US
Zibo Meng
Zibo Meng
OPPO
Computer VisionImage RestorationGenAI