🤖 AI Summary
Current instruction-based image editing methods exhibit significant performance degradation on complex scenes involving multiple entities, particularly struggling with referring expression understanding. This work identifies systematic deficiencies in mainstream models for multi-entity referring editing tasks. To address this, we introduce RefEdit-Bench—the first high-quality, referring-expression-oriented benchmark—and propose RefEdit, a lightweight and efficient model. RefEdit leverages a scalable synthetic data generation pipeline, built upon RefCOCO annotations and adapted to diffusion models (Flux/SD3), achieving superior performance with only 20K training samples—outperforming million-scale baseline methods. On RefEdit-Bench, RefEdit substantially surpasses leading closed- and open-source approaches; it also attains state-of-the-art results on conventional image editing benchmarks. All code, data, and model weights are publicly released.
📝 Abstract
Despite recent advances in inversion and instruction-based image editing, existing approaches primarily excel at editing single, prominent objects but significantly struggle when applied to complex scenes containing multiple entities. To quantify this gap, we first introduce RefEdit-Bench, a rigorous real-world benchmark rooted in RefCOCO, where even baselines trained on millions of samples perform poorly. To overcome this limitation, we introduce RefEdit -- an instruction-based editing model trained on our scalable synthetic data generation pipeline. Our RefEdit, trained on only 20,000 editing triplets, outperforms the Flux/SD3 model-based baselines trained on millions of data. Extensive evaluations across various benchmarks demonstrate that our model not only excels in referring expression tasks but also enhances performance on traditional benchmarks, achieving state-of-the-art results comparable to closed-source methods. We release data &checkpoint for reproducibility.