Multimodal Latent Diffusion Model for Complex Sewing Pattern Generation

📅 2024-12-19
🏛️ arXiv.org
📈 Citations: 6
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenges of fine-grained control and poor adaptability across diverse body shapes in complex garment pattern generation, this paper proposes a multimodal controllable sewing pattern generation method. Methodologically, it introduces a novel two-stage diffusion training paradigm that jointly conditions generation on textual descriptions, 3D human shape parameters, and hand-drawn sketches. It further designs a sewing-pattern vector expansion mechanism and a compact latent space modeling strategy to enable efficient representation and synthesis of vector-based cutting patterns. Key technical contributions include: (1) multimodal conditional embedding alignment, (2) latent space compression optimization, and (3) sewing-structure-aware vector representation learning. Experiments demonstrate that our method significantly outperforms existing approaches in style complexity, anatomical fit, and editing controllability, enabling high-fidelity, editable, and cross-body-type cutting pattern generation.

Technology Category

Application Category

📝 Abstract
Generating sewing patterns in garment design is receiving increasing attention due to its CG-friendly and flexible-editing nature. Previous sewing pattern generation methods have been able to produce exquisite clothing, but struggle to design complex garments with detailed control. To address these issues, we propose SewingLDM, a multi-modal generative model that generates sewing patterns controlled by text prompts, body shapes, and garment sketches. Initially, we extend the original vector of sewing patterns into a more comprehensive representation to cover more intricate details and then compress them into a compact latent space. To learn the sewing pattern distribution in the latent space, we design a two-step training strategy to inject the multi-modal conditions, ie, body shapes, text prompts, and garment sketches, into a diffusion model, ensuring the generated garments are body-suited and detail-controlled. Comprehensive qualitative and quantitative experiments show the effectiveness of our proposed method, significantly surpassing previous approaches in terms of complex garment design and various body adaptability. Our project page: https://shengqiliu1.github.io/SewingLDM.
Problem

Research questions and friction points this paper is trying to address.

Generating complex sewing patterns with detailed control
Integrating multi-modal conditions for garment design
Improving body adaptability in generated sewing patterns
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal generative model for sewing patterns
Two-step training strategy for condition injection
Latent space compression for intricate details
🔎 Similar Papers
No similar papers found.
Shengqi Liu
Shengqi Liu
Unknown affiliation
Y
Yuhao Cheng
MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, China
Z
Zhuo Chen
MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, China
Xingyu Ren
Xingyu Ren
Ph.D. graduate, Shanghai Jiao Tong University
Face ModelingGenerative AI
W
Wenhan Zhu
Xueshen AI
Lincheng Li
Lincheng Li
Netease Fuxi AI Lab
computer vision3D visionvideo synthesismulti-view stereo
Mengxiao Bi
Mengxiao Bi
Fuxi AI Lab, NetEase Inc.
Deep Learning
X
Xiaokang Yang
MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, China
Y
Yichao Yan
MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, China