HairShifter: Consistent and High-Fidelity Video Hair Transfer via Anchor-Guided Animation

📅 2025-07-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Video-based hair transfer faces challenges including temporal inconsistency, low spatial fidelity, and poor dynamic adaptability. This paper proposes a two-stage “anchor-frame-guided + animation-generation” framework: first, an Image Hair Transfer module achieves high-fidelity single-frame hairstyle transfer; second, a multi-scale gated SPADE decoder, augmented with semantic-aware modulation, explicitly models spatiotemporal coherence and geometric alignment within hair regions. To our knowledge, this is the first work to introduce anchor-frame guidance into video-level hair transfer—ensuring all non-hair regions remain strictly unchanged while significantly improving inter-frame coherence and fine-grained detail reconstruction. Our method achieves state-of-the-art performance across multiple benchmarks, supports diverse hairstyle transfers, and delivers superior visual quality and temporal stability. Code will be publicly released.

Technology Category

Application Category

📝 Abstract
Hair transfer is increasingly valuable across domains such as social media, gaming, advertising, and entertainment. While significant progress has been made in single-image hair transfer, video-based hair transfer remains challenging due to the need for temporal consistency, spatial fidelity, and dynamic adaptability. In this work, we propose HairShifter, a novel "Anchor Frame + Animation" framework that unifies high-quality image hair transfer with smooth and coherent video animation. At its core, HairShifter integrates a Image Hair Transfer (IHT) module for precise per-frame transformation and a Multi-Scale Gated SPADE Decoder to ensure seamless spatial blending and temporal coherence. Our method maintains hairstyle fidelity across frames while preserving non-hair regions. Extensive experiments demonstrate that HairShifter achieves state-of-the-art performance in video hairstyle transfer, combining superior visual quality, temporal consistency, and scalability. The code will be publicly available. We believe this work will open new avenues for video-based hairstyle transfer and establish a robust baseline in this field.
Problem

Research questions and friction points this paper is trying to address.

Achieving temporal consistency in video hair transfer
Maintaining high spatial fidelity during hairstyle transfer
Ensuring dynamic adaptability across video frames
Innovation

Methods, ideas, or system contributions that make the work stand out.

Anchor Frame + Animation framework
Multi-Scale Gated SPADE Decoder
Image Hair Transfer module
🔎 Similar Papers
No similar papers found.
W
Wangzheng Shi
Xiamen University, School of Informatics
Yinglin Zheng
Yinglin Zheng
Xiamen University
Computer Vision
Y
Yuxin Lin
Xiamen University, School of Informatics
Jianmin Bao
Jianmin Bao
Microsoft Research
Computer VisionAIGCDeep Generative ModelsDeep Learning
M
Ming Zeng
Xiamen University, School of Informatics
D
Dong Chen
Microsoft Research Asia, Beijing, China