🤖 AI Summary
While generative image super-resolution (SR) models enhance visual fidelity, they often introduce artifacts with varying perceptual saliency—yet existing evaluation methods treat all artifacts as binary defects, ignoring human visual sensitivity differences. Method: We propose the novel concept of “artifact saliency” and construct the first SR artifact dataset (1,302 samples), annotated via crowdsourcing with fine-grained saliency scores. Leveraging this dataset, we design a lightweight regression network that generates pixel-level artifact saliency heatmaps. Contribution/Results: Experiments demonstrate that our method outperforms conventional metrics in localizing perceptually salient artifacts, and the generated heatmaps closely align with human visual attention regions. This work shifts SR evaluation from global quality assessment toward fine-grained perceptual modeling. To foster future research, we publicly release both the dataset and code, establishing a new benchmark for perception-driven SR optimization.
📝 Abstract
Generative image super-resolution (SR) is rapidly advancing in visual quality and detail restoration. As the capacity of SR models expands, however, so does their tendency to produce artifacts: incorrect, visually disturbing details that reduce perceived quality. Crucially, their perceptual impact varies: some artifacts are barely noticeable while others strongly degrade the image. We argue that artifacts should be characterized by their prominence to human observers rather than treated as uniform binary defects. Motivated by this, we present a novel dataset of 1302 artifact examples from 11 contemporary image-SR methods, where each artifact is paired with a crowdsourced prominence score. Building on this dataset, we train a lightweight regressor that produces spatial prominence heatmaps and outperforms existing methods at detecting prominent artifacts. We release the dataset and code to facilitate prominence-aware evaluation and mitigation of SR artifacts.