A Unified Framework for Multimodal Image Reconstruction and Synthesis using Denoising Diffusion Models

📅 2026-02-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing approaches require separate, task-specific models for multimodal image reconstruction and synthesis, leading to complex pipelines and limited generalization. This work proposes Any2all, a unified framework that, for the first time, formulates both tasks as a virtual image inpainting problem. By training a single unconditional diffusion model on complete multimodal data, the framework enables flexible inference with arbitrary input combinations to generate target modality images. Without any task-specific customization, Any2all achieves high-fidelity reconstruction and synthesis simultaneously on PET/MR/CT brain datasets, significantly outperforming specialized models and attaining state-of-the-art performance in both distortion metrics and perceptual quality.

Technology Category

Application Category

📝 Abstract
Image reconstruction and image synthesis are important for handling incomplete multimodal imaging data, but existing methods require various task-specific models, complicating training and deployment workflows. We introduce Any2all, a unified framework that addresses this limitation by formulating these disparate tasks as a single virtual inpainting problem. We train a single, unconditional diffusion model on the complete multimodal data stack. This model is then adapted at inference time to ``inpaint''all target modalities from any combination of inputs of available clean images or noisy measurements. We validated Any2all on a PET/MR/CT brain dataset. Our results show that Any2all can achieve excellent performance on both multimodal reconstruction and synthesis tasks, consistently yielding images with competitive distortion-based performance and superior perceptual quality over specialized methods.
Problem

Research questions and friction points this paper is trying to address.

multimodal image reconstruction
image synthesis
incomplete multimodal imaging data
task-specific models
medical imaging
Innovation

Methods, ideas, or system contributions that make the work stand out.

unified framework
denoising diffusion models
multimodal image reconstruction
image synthesis
virtual inpainting
🔎 Similar Papers
No similar papers found.
Weijie Gan
Weijie Gan
PhD in CSE @ WashU
Machine LearningComputer VisionComputational Imaging
X
Xucheng Wang
Department of Computer Science and Engineering, Washington University in St. Louis, St. Louis, MO, USA
T
Tongyao Wang
Mallinckrodt Institute of Radiology, Washington University in St. Louis, St. Louis, MO USA
W
Wenshang Wang
Mallinckrodt Institute of Radiology, Washington University in St. Louis, St. Louis, MO USA
C
Chunwei Ying
Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, USA
Yuyang Hu
Yuyang Hu
PhD student in ESE, Washington university in St. Louis
Computational ImagingInverse ProblemsMachine LearningSignal Processing
Y
Yasheng Chen
Department of Neurology, Washington University in St. Louis, St. Louis, MO USA
Hongyu An
Hongyu An
Washington University in St. Louis
MR imagingPET/MRCerebrovascular Disease Imaging
Ulugbek S. Kamilov
Ulugbek S. Kamilov
Donald L. Snyder Associate Professor of ESE and CSE at WashU
Computational ImagingBiomedical ImagingMachine LearningOptimizationSignal Processing