🤖 AI Summary
Existing video quality assessment (VQA) methods typically rely on single-model architectures producing scalar scores, suffering from poor interpretability, limited cross-format generalization, and inability to diagnose specific distortion types. To address these limitations, we propose the Diagnostic Mixture-of-Experts (Diag-MoE) framework: a multi-agent expert architecture trained collaboratively without fine-tuning to achieve universal modeling; it incorporates pairwise ranking loss, multi-task learning heads, and weak supervision to jointly optimize global quality scores and multidimensional artifact vectors on a large-scale, self-constructed database. Extensive evaluation across 17 diverse benchmarks—including HD, UHD, HDR, and HFR datasets—demonstrates that Diag-MoE consistently outperforms 18 state-of-the-art baselines under static parameter settings. To our knowledge, this is the first VQA method achieving high accuracy, intrinsic interpretability, and true universality across heterogeneous video formats and distortions.
📝 Abstract
Recent works in video quality assessment (VQA) typically employ monolithic models that typically predict a single quality score for each test video. These approaches cannot provide diagnostic, interpretable feedback, offering little insight into why the video quality is degraded. Most of them are also specialized, format-specific metrics rather than truly ``generic"solutions, as they are designed to learn a compromised representation from disparate perceptual domains. To address these limitations, this paper proposes Unified-VQA, a framework that provides a single, unified quality model applicable to various distortion types within multiple video formats by recasting generic VQA as a Diagnostic Mixture-of-Experts (MoE) problem. Unified-VQA employs multiple ``perceptual experts''dedicated to distinct perceptual domains. A novel multi-proxy expert training strategy is designed to optimize each expert using a ranking-inspired loss, guided by the most suitable proxy metric for its domain. We also integrated a diagnostic multi-task head into this framework to generate a global quality score and an interpretable multi-dimensional artifact vector, which is optimized using a weakly-supervised learning strategy, leveraging the known properties of the large-scale training database generated for this work. With static model parameters (without retraining or fine-tuning), Unified-VQA demonstrates consistent and superior performance compared to over 18 benchmark methods for both generic VQA and diagnostic artifact detection tasks across 17 databases containing diverse streaming artifacts in HD, UHD, HDR and HFR formats. This work represents an important step towards practical, actionable, and interpretable video quality assessment.