A Modality-agnostic Multi-task Foundation Model for Human Brain Imaging

📅 2025-08-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing medical image calibration methods perform well on CT but exhibit poor generalizability to uncalibrated multi-protocol MRI (e.g., T1w/T2w/FLAIR), being highly sensitive to variations in contrast, resolution, and acquisition orientation. To address this, we propose BrainFM—the first modality-agnostic, multitask vision foundation model for brain imaging. BrainFM employs a “mild-to-severe” intra-subject image generation scheme and a “real-synthetic” hybrid training strategy, augmented by anatomical prior guidance, contextual modeling, and a unified multitask architecture, thereby substantially enhancing robustness to appearance variability. Without task-specific fine-tuning, BrainFM simultaneously supports cross-modal synthesis (MRI/CT), segmentation, and registration. Evaluated across 11 public benchmarks, it consistently outperforms state-of-the-art single-task models, demonstrates stable cross-protocol performance, and establishes a scalable foundation model paradigm for clinical heterogeneous neuroimaging analysis.

Technology Category

Application Category

📝 Abstract
Recent learning-based approaches have made astonishing advances in calibrated medical imaging like computerized tomography (CT), yet they struggle to generalize in uncalibrated modalities -- notably magnetic resonance (MR) imaging, where performance is highly sensitive to the differences in MR contrast, resolution, and orientation. This prevents broad applicability to diverse real-world clinical protocols. Here we introduce BrainFM, a modality-agnostic, multi-task vision foundation model for human brain imaging. With the proposed "mild-to-severe" intra-subject generation and "real-synth" mix-up training strategy, BrainFM is resilient to the appearance of acquired images (e.g., modality, contrast, deformation, resolution, artifacts), and can be directly applied to five fundamental brain imaging tasks, including image synthesis for CT and T1w/T2w/FLAIR MRI, anatomy segmentation, scalp-to-cortical distance, bias field estimation, and registration. We evaluate the efficacy of BrainFM on eleven public datasets, and demonstrate its robustness and effectiveness across all tasks and input modalities. Code is available at https://github.com/jhuldr/BrainFM.
Problem

Research questions and friction points this paper is trying to address.

Generalizing across uncalibrated MR imaging modalities
Handling appearance variations in brain imaging data
Enabling broad applicability to diverse clinical protocols
Innovation

Methods, ideas, or system contributions that make the work stand out.

Modality-agnostic multi-task foundation model
Intra-subject generation and mix-up training
Resilient to diverse image appearances
🔎 Similar Papers
No similar papers found.
Peirong Liu
Peirong Liu
Assistant Professor of ECE, Johns Hopkins University
AI for HealthcareComputer VisionMedical Imaging
Oula Puonti
Oula Puonti
Danish Research Centre for Magnetic Resonance
Bayesian modelingMedical Image AnalysisNon-invasive brain stimulation
X
Xiaoling Hu
Athinoula A. Martinos Center for Biomedical Imaging, Harvard Medical School and Massachusetts General Hospital, Boston, MA 02129, USA
Karthik Gopinath
Karthik Gopinath
Postdoctoral Research Fellow @ Massachusetts General Hospital & Harvard Medical School
Medical Image AnalysisMachine Learning
Annabel Sorby-Adams
Annabel Sorby-Adams
Harvard Medical School and Massachusetts General Hospital
Daniel C. Alexander
Daniel C. Alexander
Professor of Imaging Science, Centre for Medical Image Computing, Department of Computer Science
Computer scienceMachine learningMedical imagingdiffusion MRINeuroscience
W
W. Taylor Kimberly
Athinoula A. Martinos Center for Biomedical Imaging, Harvard Medical School and Massachusetts General Hospital, Boston, MA 02129, USA
J
Juan E. Iglesias
Athinoula A. Martinos Center for Biomedical Imaging, Harvard Medical School and Massachusetts General Hospital, Boston, MA 02129, USA