Uncertainty-Aware ControlNet: Bridging Domain Gaps with Synthetic Image Generation

📅 2025-10-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
ControlNet often replicates the source-domain distribution in cross-domain image synthesis, struggling to adapt to unlabeled target domains (e.g., low-quality Home-OCT OCT images). To address this, we propose Uncertainty-ControlNet—the first method to embed uncertainty modeling directly into ControlNet’s control mechanism. It jointly leverages semantic guidance from labeled source data and uncertainty-driven guidance from unlabeled target-domain inputs, enabling arbitrary-domain image generation without explicit style alignment. Our model integrates a diffusion-based architecture, dual-control inputs (semantic + uncertainty), and automatic synthetic annotation of generated images, supporting unsupervised synthesis of labeled data for the target domain. Evaluated on retinal OCT and traffic-scene cross-domain segmentation tasks, Uncertainty-ControlNet achieves effective domain adaptation using only synthetically labeled target-domain data—outperforming style-transfer-based and other baseline methods by significant margins in segmentation accuracy.

Technology Category

Application Category

📝 Abstract
Generative Models are a valuable tool for the controlled creation of high-quality image data. Controlled diffusion models like the ControlNet have allowed the creation of labeled distributions. Such synthetic datasets can augment the original training distribution when discriminative models, like semantic segmentation, are trained. However, this augmentation effect is limited since ControlNets tend to reproduce the original training distribution. This work introduces a method to utilize data from unlabeled domains to train ControlNets by introducing the concept of uncertainty into the control mechanism. The uncertainty indicates that a given image was not part of the training distribution of a downstream task, e.g., segmentation. Thus, two types of control are engaged in the final network: an uncertainty control from an unlabeled dataset and a semantic control from the labeled dataset. The resulting ControlNet allows us to create annotated data with high uncertainty from the target domain, i.e., synthetic data from the unlabeled distribution with labels. In our scenario, we consider retinal OCTs, where typically high-quality Spectralis images are available with given ground truth segmentations, enabling the training of segmentation networks. The recent development in Home-OCT devices, however, yields retinal OCTs with lower quality and a large domain shift, such that out-of-the-pocket segmentation networks cannot be applied for this type of data. Synthesizing annotated images from the Home-OCT domain using the proposed approach closes this gap and leads to significantly improved segmentation results without adding any further supervision. The advantage of uncertainty-guidance becomes obvious when compared to style transfer: it enables arbitrary domain shifts without any strict learning of an image style. This is also demonstrated in a traffic scene experiment.
Problem

Research questions and friction points this paper is trying to address.

Bridging domain gaps in image data using synthetic generation
Improving segmentation on unlabeled domains with uncertainty control
Generating annotated synthetic data for domain adaptation tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Introducing uncertainty control into ControlNet mechanism
Generating synthetic data from unlabeled target domains
Enabling domain adaptation without additional supervision requirements
🔎 Similar Papers
No similar papers found.
J
Joshua Niemeijer
German Aerospace Center (DLR)
J
Jan Ehrhardt
University of Lübbeck
Heinz Handels
Heinz Handels
Professor of Medical Informatics, Director DFKI, University of Lübeck
Medical Image ComputingArtificial IntelligenceDeep LearningVirtual Reality Simulations
H
Hristina Uzunova
German Research Center for Artificial Intelligence (DFKI)