Any-to-All MRI Synthesis: A Unified Foundation Model for Nasopharyngeal Carcinoma and Its Downstream Applications

📅 2026-02-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of missing multimodal MRI in nasopharyngeal carcinoma (NPC) radiotherapy planning, often caused by patient discomfort, high scanning costs, and prolonged acquisition times. To this end, the authors propose the first unified any-to-all MRI synthesis framework tailored for NPC radiotherapy. By leveraging contrastive learning to extract modality-invariant representations and integrating a CLIP-driven text-guided decoder, the model enables semantically consistent and controllable synthesis of all target MRI modalities. Trained on 40,825 images from 13 institutions and validated across 26 internal and external test sets comprising 15,748 images, the framework achieves an average SSIM of 0.90 and PSNR of 27, significantly improving synthesis fidelity and robustness. Moreover, it enhances performance in downstream radiotherapy tasks such as segmentation, while demonstrating strong anatomical adaptability, clinical interpretability, and cross-modal generalization capability.

Technology Category

Application Category

📝 Abstract
Magnetic resonance imaging (MRI) is essential for nasopharyngeal carcinoma (NPC) radiotherapy (RT), but practical constraints, such as patient discomfort, long scan times, and high costs often lead to incomplete modalities in clinical practice, compromising RT planning accuracy. Traditional MRI synthesis methods are modality-specific, limited in anatomical adaptability, and lack clinical interpretability-failing to meet NPC's RT needs. Here, we developed a unified foundation model integrating contrastive visual representation learning and vision-language alignment (VLA) to enable any-to-all MRI synthesis. The model uses a contrastive encoder for modality-invariant representations and a CLIP-based text-informed decoder for semantically consistent synthesis, supporting any-to-all MRI synthesis via one unified foundation model. Trained on 40,825 images from 13 institutions, it achieves consistently high performance (average SSIM 0.90, PSNR 27) across 26 internal/external validation sites (15,748 images), with superior synthesis fidelity and robustness to noise and domain shifts. Meanwhile, its unified representation enhances downstream RT-relevant tasks (e.g., segmentation). This work advances digital medicine solutions for NPC care by leveraging foundation models to bridge technical synthesis and clinical utility.
Problem

Research questions and friction points this paper is trying to address.

MRI synthesis
nasopharyngeal carcinoma
radiotherapy planning
incomplete modalities
clinical interpretability
Innovation

Methods, ideas, or system contributions that make the work stand out.

any-to-all MRI synthesis
foundation model
contrastive representation learning
vision-language alignment
nasopharyngeal carcinoma
🔎 Similar Papers
No similar papers found.
Y
Yao Pu
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
Yiming Shi
Yiming Shi
University of Electronic Science and Technology of China
Efficient AIParameter Efficient Fine TuningDiffusionMultimodal
Z
Zhenxi Zhang
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
P
Peixin Yu
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
Y
Yitao Zhuang
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
Xiang Wang
Xiang Wang
University of Science and Technology of China
Trustworthy AIGraph LearningRecommendationFoundation ModelsMultimodal Models
H
Hongzhao Chen
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
J
Jing Cai
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
Ge Ren
Ge Ren
Shanghai Jiao Tong University
AI Intellectual Property Protection