FetalFlex: Anatomy-Guided Diffusion Model for Flexible Control on Fetal Ultrasound Image Synthesis

📅 2025-03-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Fetal ultrasound multiplanar annotation data—particularly for rare anomalies—are severely scarce, hindering AI model training and clinical education. To address this, we propose an anatomy-guided controllable diffusion generation framework that enables out-of-distribution fetal ultrasound image synthesis without requiring real abnormal samples. Our method integrates anatomical prior encoding, a pre-alignment module, Repaint-based texture-consistent inpainting, and a two-stage adaptive sampling strategy to ensure cross-plane anatomical consistency and pathologically plausible abnormalities. Evaluated on multicenter datasets, our approach achieves state-of-the-art image quality (FID reduced by 23.6%, LPIPS reduced by 18.4%) and significantly improves downstream anomaly detection across six models (average AUC increased by 5.2%). Clinical credibility is further validated through blinded expert radiologist assessment.

Technology Category

Application Category

📝 Abstract
Fetal ultrasound (US) examinations require the acquisition of multiple planes, each providing unique diagnostic information to evaluate fetal development and screening for congenital anomalies. However, obtaining a comprehensive, multi-plane annotated fetal US dataset remains challenging, particularly for rare or complex anomalies owing to their low incidence and numerous subtypes. This poses difficulties in training novice radiologists and developing robust AI models, especially for detecting abnormal fetuses. In this study, we introduce a Flexible Fetal US image generation framework (FetalFlex) to address these challenges, which leverages anatomical structures and multimodal information to enable controllable synthesis of fetal US images across diverse planes. Specifically, FetalFlex incorporates a pre-alignment module to enhance controllability and introduces a repaint strategy to ensure consistent texture and appearance. Moreover, a two-stage adaptive sampling strategy is developed to progressively refine image quality from coarse to fine levels. We believe that FetalFlex is the first method capable of generating both in-distribution normal and out-of-distribution abnormal fetal US images, without requiring any abnormal data. Experiments on multi-center datasets demonstrate that FetalFlex achieved state-of-the-art performance across multiple image quality metrics. A reader study further confirms the close alignment of the generated results with expert visual assessments. Furthermore, synthetic images by FetalFlex significantly improve the performance of six typical deep models in downstream classification and anomaly detection tasks. Lastly, FetalFlex's anatomy-level controllable generation offers a unique advantage for anomaly simulation and creating paired or counterfactual data at the pixel level. The demo is available at: https://dyf1023.github.io/FetalFlex/.
Problem

Research questions and friction points this paper is trying to address.

Generates diverse fetal ultrasound images for training and AI development.
Enhances image synthesis control using anatomical structures and multimodal data.
Improves anomaly detection and classification in fetal ultrasound imaging.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Anatomy-guided diffusion model for fetal US synthesis
Pre-alignment module enhances image controllability
Two-stage adaptive sampling refines image quality
🔎 Similar Papers
No similar papers found.
Y
Yaofei Duan
Faculty of Applied Sciences, Macao Polytechnic University, Macao, China
Tao Tan
Tao Tan
FCA MPU
Medical Imaging AI
Zhiyuan Zhu
Zhiyuan Zhu
Shanghai Jiao Tong University
NLPASRTTS
Yuhao Huang
Yuhao Huang
Shenzhen University
Medical Image ComputingUltrasoundModel Robustness
Yuanji Zhang
Yuanji Zhang
Shenzhen University
Intelligent ultrasoundFetal growth and developmentFetal malformations
R
Rui Gao
Shenzhen RayShape Medical Technology Co., Ltd, Shenzhen, Guangdong, China
Patrick Cheong-Iao Pang
Patrick Cheong-Iao Pang
Faculty of Applied Sciences, Macao Polytechnic University
human-computer interactioneducational technologyAI for sustainability
X
Xinru Gao
Northwest Women & Children Hospital, Xian, Shaanxi, China
Guowei Tao
Guowei Tao
Qilu Hospital of Shandong University, Jinan, Shandong, China
X
Xiang Cong
Qilu Hospital of Shandong University, Jinan, Shandong, China
Z
Zhou Li
Department of Ultrasound, Shenzhen Guangming District People’s Hospital, Shenzhen, Guangdong, China; Jinan University, Guangzhou, Guangdong, China
L
Lianying Liang
Department of Ultrasound, Shenzhen Guangming District People’s Hospital, Shenzhen, Guangdong, China; Jinan University, Guangzhou, Guangdong, China
G
Guangzhi He
Department of Ultrasound, Shenzhen Guangming District People’s Hospital, Shenzhen, Guangdong, China; Jinan University, Guangzhou, Guangdong, China
L
Linliang Yin
Center for Medical Ultrasound, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou Municipal Hospital, Gusu School, Nanjing Medical University, Suzhou, Jiangsu, China
X
Xuedong Deng
Center for Medical Ultrasound, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou Municipal Hospital, Gusu School, Nanjing Medical University, Suzhou, Jiangsu, China
X
Xin Yang
National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, Guangdong, China; Medical Ultrasound Image Computing (MUSIC) Laboratory, Shenzhen University, Shenzhen, Guangdong, China
D
Dong Ni
National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, Guangdong, China; Medical Ultrasound Image Computing (MUSIC) Laboratory, Shenzhen University, Shenzhen, Guangdong, China