🤖 AI Summary
Current text-to-image models exhibit limited capability in aligning generated images with complex textual descriptions involving multiple instances, diverse categories, and intricate semantic relationships; moreover, fine-grained evaluation benchmarks strongly correlated with human judgment remain scarce. To address these challenges, we introduce M3R-Bench—the first large-scale, multi-category, multi-instance, multi-relation image-text alignment benchmark—and propose Revise-Then-Enforce, a training-free post-editing method. We further design AlignScore, an automatic metric that jointly models object detection and semantic parsing to assess both visual entities and their relational structure, achieving strong correlation with human preferences (ρ > 0.85). Experiments reveal significant performance bottlenecks of mainstream open-source diffusion models on M3R-Bench, confirming its rigor; Revise-Then-Enforce consistently improves alignment quality across models—including Stable Diffusion—yielding an average AlignScore gain of 12.7%. This work establishes a new paradigm for evaluating and optimizing image-text alignment.
📝 Abstract
Text-to-image models are known to struggle with generating images that perfectly align with textual prompts. Several previous studies have focused on evaluating image-text alignment in text-to-image generation. However, these evaluations either address overly simple scenarios, especially overlooking the difficulty of prompts with multiple different instances belonging to the same category, or they introduce metrics that do not correlate well with human evaluation. In this study, we introduce M$^3$T2IBench, a large-scale, multi-category, multi-instance, multi-relation along with an object-detection-based evaluation metric, $AlignScore$, which aligns closely with human evaluation. Our findings reveal that current open-source text-to-image models perform poorly on this challenging benchmark. Additionally, we propose the Revise-Then-Enforce approach to enhance image-text alignment. This training-free post-editing method demonstrates improvements in image-text alignment across a broad range of diffusion models. footnote{Our code and data has been released in supplementary material and will be made publicly available after the paper is accepted.}