🤖 AI Summary
The rapid advancement of generative AI image models has rendered traditional human-annotated quality assessment approaches inadequate. To address this challenge, this work proposes ELIQ, a novel framework that enables the first unsupervised, label-free evaluation focusing on visual quality and text-image alignment. Built upon pretrained multimodal models, ELIQ leverages instruction tuning, lightweight gated fusion, and a Quality Query Transformer to automatically construct positive-negative sample pairs and train a dual-dimension quality predictor. Extensive experiments demonstrate that ELIQ significantly outperforms existing label-free methods across multiple benchmarks, exhibits strong generalization across model generations, and seamlessly transfers to user-generated content scenarios, offering a scalable solution for evaluating dynamically evolving generative models.
📝 Abstract
Generative text-to-image models are advancing at an unprecedented pace, continuously shifting the perceptual quality ceiling and rendering previously collected labels unreliable for newer generations. To address this, we present ELIQ, a Label-free Framework for Quality Assessment of Evolving AI-generated Images. Specifically, ELIQ focuses on visual quality and prompt-image alignment, automatically constructs positive and aspect-specific negative pairs to cover both conventional distortions and AIGC-specific distortion modes, enabling transferable supervision without human annotations. Building on these pairs, ELIQ adapts a pre-trained multimodal model into a quality-aware critic via instruction tuning and predicts two-dimensional quality using lightweight gated fusion and a Quality Query Transformer. Experiments across multiple benchmarks demonstrate that ELIQ consistently outperforms existing label-free methods, generalizes from AI-generated content (AIGC) to user-generated content (UGC) scenarios without modification, and paves the way for scalable and label-free quality assessment under continuously evolving generative models. The code will be released upon publication.