🤖 AI Summary
Existing image customization methods are predominantly single-task oriented, lacking the flexibility to jointly control and combine diverse conditions—such as identity, subject, style, and background. To address this, we propose UniCustom, the first unified image customization framework. It introduces two key innovations: feature-routing constraints for disentangled semantic control and position-aware placeholder tokens for precise spatial conditioning, both integrated within a Diffusion Transformer (DiT) architecture. A three-stage progressive training strategy further enables accurate multi-condition decoupling and high-fidelity generation. UniCustom is trained end-to-end on a large-scale, multi-task dataset with joint optimization, significantly improving generalization across unseen condition combinations and output consistency. Extensive experiments demonstrate that UniCustom achieves state-of-the-art fidelity and fine-grained controllability across diverse customization tasks, establishing a scalable, general-purpose paradigm for conditional image generation.
📝 Abstract
Recently, extensive research on image customization (e.g., identity, subject, style, background, etc.) demonstrates strong customization capabilities in large-scale generative models. However, most approaches are designed for specific tasks, restricting their generalizability to combine different types of condition. Developing a unified framework for image customization remains an open challenge. In this paper, we present DreamO, an image customization framework designed to support a wide range of tasks while facilitating seamless integration of multiple conditions. Specifically, DreamO utilizes a diffusion transformer (DiT) framework to uniformly process input of different types. During training, we construct a large-scale training dataset that includes various customization tasks, and we introduce a feature routing constraint to facilitate the precise querying of relevant information from reference images. Additionally, we design a placeholder strategy that associates specific placeholders with conditions at particular positions, enabling control over the placement of conditions in the generated results. Moreover, we employ a progressive training strategy consisting of three stages: an initial stage focused on simple tasks with limited data to establish baseline consistency, a full-scale training stage to comprehensively enhance the customization capabilities, and a final quality alignment stage to correct quality biases introduced by low-quality data. Extensive experiments demonstrate that the proposed DreamO can effectively perform various image customization tasks with high quality and flexibly integrate different types of control conditions.