MDIQA: Unified Image Quality Assessment for Multi-dimensional Evaluation and Restoration

📅 2025-08-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing image quality assessment (IQA) methods predominantly produce a single holistic score, overlooking the multi-dimensional nature of human perception—encompassing both technical and aesthetic attributes. To address this limitation, we propose MDIQA, a novel multidimensional IQA framework that jointly models five technical dimensions (e.g., noise, blur) and four aesthetic dimensions (e.g., composition, color harmony). MDIQA employs a multi-branch deep network to extract dimension-specific features, followed by supervised feature fusion and jointly optimized, adjustable dimension weights. Crucially, it enables user-preference-driven training for customizable image restoration. Extensive experiments demonstrate that MDIQA achieves state-of-the-art performance across multiple mainstream IQA benchmarks. Moreover, when integrated into image restoration pipelines, it significantly enhances subjective visual quality, validating its effectiveness in real-world, preference-aware applications.

Technology Category

Application Category

📝 Abstract
Recent advancements in image quality assessment (IQA), driven by sophisticated deep neural network designs, have significantly improved the ability to approach human perceptions. However, most existing methods are obsessed with fitting the overall score, neglecting the fact that humans typically evaluate image quality from different dimensions before arriving at an overall quality assessment. To overcome this problem, we propose a multi-dimensional image quality assessment (MDIQA) framework. Specifically, we model image quality across various perceptual dimensions, including five technical and four aesthetic dimensions, to capture the multifaceted nature of human visual perception within distinct branches. Each branch of our MDIQA is initially trained under the guidance of a separate dimension, and the respective features are then amalgamated to generate the final IQA score. Additionally, when the MDIQA model is ready, we can deploy it for a flexible training of image restoration (IR) models, enabling the restoration results to better align with varying user preferences through the adjustment of perceptual dimension weights. Extensive experiments demonstrate that our MDIQA achieves superior performance and can be effectively and flexibly applied to image restoration tasks. The code is available: https://github.com/YaoShunyu19/MDIQA.
Problem

Research questions and friction points this paper is trying to address.

Assesses image quality across multiple perceptual dimensions
Overcomes neglect of human multidimensional evaluation in IQA
Enables flexible image restoration aligned with user preferences
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-dimensional image quality assessment framework
Separate training for technical and aesthetic dimensions
Flexible image restoration via adjustable dimension weights
🔎 Similar Papers
No similar papers found.
S
Shunyu Yao
Harbin Institute of Technology
M
Ming Liu
Harbin Institute of Technology
Zhilu Zhang
Zhilu Zhang
Harbin Institute of Technology
Low-Level VisionComputational Photography3D Reconstruction and Generation
Z
Zhaolin Wan
Harbin Institute of Technology
Z
Zhilong Ji
Tomorrow Advancing Life
J
Jinfeng Bai
Tomorrow Advancing Life
Wangmeng Zuo
Wangmeng Zuo
School of Computer Science and Technology, Harbin Institute of Technology
Computer VisionImage ProcessingGenerative AIDeep LearningBiometrics