Prototypicality Bias Reveals Blindspots in Multimodal Evaluation Metrics

📅 2026-01-08
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the susceptibility of existing automatic evaluation metrics for text-to-image generation to prototype bias, which often favors visually or socially typical yet semantically incorrect images. To systematically assess the ability of multimodal metrics to distinguish between prototypicality and semantic fidelity, the authors introduce ProtoBias, a contrastive benchmark. They propose ProtoScore, an efficient and robust metric based on a custom 7B-parameter model, and conduct comprehensive analyses using controlled datasets alongside CLIPScore, PickScore, VQA, and LLM-as-Judge approaches. Experiments reveal that prevailing metrics consistently exhibit prototype bias, whereas human evaluators prioritize semantic correctness. ProtoScore significantly reduces misjudgment rates while maintaining high inference speed, achieving robustness comparable to large closed-source models.

Technology Category

Application Category

📝 Abstract
Automatic metrics are now central to evaluating text-to-image models, often substituting for human judgment in benchmarking and large-scale filtering. However, it remains unclear whether these metrics truly prioritize semantic correctness or instead favor visually and socially prototypical images learned from biased data distributions. We identify and study prototypicality bias as a systematic failure mode in multimodal evaluation. We introduce a controlled contrastive benchmark ProtoBias (Prototypical Bias), spanning Animals, Objects, and Demography images, where semantically correct but non-prototypical images are paired with subtly incorrect yet prototypical adversarial counterparts. This setup enables a directional evaluation of whether metrics follow textual semantics or default to prototypes. Our results show that widely used metrics, including CLIPScore, PickScore, and VQA-based scores, frequently misrank these pairs, while even LLM-as-Judge systems exhibit uneven robustness in socially grounded cases. Human evaluations consistently favour semantic correctness with larger decision margins. Motivated by these findings, we propose ProtoScore, a robust 7B-parameter metric that substantially reduces failure rates and suppresses misranking, while running at orders of magnitude faster than the inference time of GPT-5, approaching the robustness of much larger closed-source judges.
Problem

Research questions and friction points this paper is trying to address.

prototypicality bias
multimodal evaluation
text-to-image generation
semantic correctness
evaluation metrics
Innovation

Methods, ideas, or system contributions that make the work stand out.

prototypicality bias
multimodal evaluation
ProtoBias
ProtoScore
text-to-image generation
🔎 Similar Papers
No similar papers found.