Med-MMFL: A Multimodal Federated Learning Benchmark in Healthcare

📅 2026-02-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the lack of a unified benchmark for medical multimodal federated learning (MMFL) that supports multimodal, multitask, and realistic heterogeneous settings, which has hindered systematic research in this area. To bridge this gap, we introduce Med-MMFL, the first comprehensive benchmark tailored to the medical domain, encompassing 2–4 modalities—including histopathology images, X-rays, ECG, MRI, and clinical text—across ten diverse medical datasets and multiple federated configurations (naturally partitioned, synthetic IID/non-IID). We evaluate six state-of-the-art federated algorithms on tasks such as segmentation, classification, modality alignment, and visual question answering, integrating various aggregation strategies, loss functions, and regularization techniques. The entire pipeline—including data preprocessing, partitioning, and training—is open-sourced to provide a reproducible and standardized evaluation platform for MMFL research.

Technology Category

Application Category

📝 Abstract
Federated learning (FL) enables collaborative model training across decentralized medical institutions while preserving data privacy. However, medical FL benchmarks remain scarce, with existing efforts focusing mainly on unimodal or bimodal modalities and a limited range of medical tasks. This gap underscores the need for standardized evaluation to advance systematic understanding in medical MultiModal FL (MMFL). To this end, we introduce Med-MMFL, the first comprehensive MMFL benchmark for the medical domain, encompassing diverse modalities, tasks, and federation scenarios. Our benchmark evaluates six representative state-of-the-art FL algorithms, covering different aggregation strategies, loss formulations, and regularization techniques. It spans datasets with 2 to 4 modalities, comprising a total of 10 unique medical modalities, including text, pathology images, ECG, X-ray, radiology reports, and multiple MRI sequences. Experiments are conducted across naturally federated, synthetic IID, and synthetic non-IID settings to simulate real-world heterogeneity. We assess segmentation, classification, modality alignment (retrieval), and VQA tasks. To support reproducibility and fair comparison of future multimodal federated learning (MMFL) methods under realistic medical settings, we release the complete benchmark implementation, including data processing and partitioning pipelines, at https://github.com/bhattarailab/Med-MMFL-Benchmark .
Problem

Research questions and friction points this paper is trying to address.

Multimodal Federated Learning
Medical Benchmark
Data Privacy
Federated Learning
Healthcare
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal Federated Learning
Medical Benchmark
Data Heterogeneity
Federated Aggregation
Cross-modal Alignment
🔎 Similar Papers
No similar papers found.
A
Aavash Chhetri
NepAl Applied Mathematics and Informatics Institute for research, Nepal
B
Bibek Niroula
NepAl Applied Mathematics and Informatics Institute for research, Nepal
P
Pratik Shrestha
NepAl Applied Mathematics and Informatics Institute for research, Nepal
Yash Raj Shrestha
Yash Raj Shrestha
University of Lausanne, Applied AI Lab
Applied AIHuman - AI CollaborationData Driven Decision-MakingOrganization Design
L
Lesley A Anderson
University of Aberdeen, Aberdeen, UK
Prashnna K Gyawali
Prashnna K Gyawali
West Virginia University
Machine LearningDeep Learning
Loris Bazzani
Loris Bazzani
Principal Computer Vision Scientist, Amazon
Computer VisionMachine LearningArtificial Intelligence
Binod Bhattarai
Binod Bhattarai
Assistant Professor, University of Aberdeen
Machine LearningMedical Image AnalysisComputer Vision