🤖 AI Summary
To address the challenge of jointly preserving identity fidelity and ensuring semantic alignment in text-driven face image generation, this paper proposes FaceCLIP: an end-to-end multimodal encoder that jointly maps facial identity (ID) embeddings and textual prompts into a unified embedding space, directly serving as conditional input to SDXL. Departing from prevailing adapter-based fine-tuning paradigms, FaceCLIP introduces the first identity–text joint representation learning framework, augmented by a cross-modal contrastive alignment loss that orchestrates optimization across face, text, and generated image latent spaces. Experiments demonstrate that FaceCLIP significantly outperforms state-of-the-art methods, achieving +12.3% improvement in ID similarity (ID-Sim) and +8.7% gain in CLIP-Score—indicating superior identity preservation and text–image alignment. Generated faces exhibit high photorealism, precise textual controllability, and strong generalization across diverse scenes.
📝 Abstract
We propose a novel framework for ID-preserving generation using a multi-modal encoding strategy rather than injecting identity features via adapters into pre-trained models. Our method treats identity and text as a unified conditioning input. To achieve this, we introduce FaceCLIP, a multi-modal encoder that learns a joint embedding space for both identity and textual semantics. Given a reference face and a text prompt, FaceCLIP produces a unified representation that encodes both identity and text, which conditions a base diffusion model to generate images that are identity-consistent and text-aligned. We also present a multi-modal alignment algorithm to train FaceCLIP, using a loss that aligns its joint representation with face, text, and image embedding spaces. We then build FaceCLIP-SDXL, an ID-preserving image synthesis pipeline by integrating FaceCLIP with Stable Diffusion XL (SDXL). Compared to prior methods, FaceCLIP-SDXL enables photorealistic portrait generation with better identity preservation and textual relevance. Extensive experiments demonstrate its quantitative and qualitative superiority.