dFLMoE: Decentralized Federated Learning via Mixture of Experts for Medical Data Analysis

📅 2025-03-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address knowledge degradation, single-point server dependency, and poor adaptability to heterogeneous data in centralized medical federated learning, this paper proposes the first decentralized Mixture of Experts (MoE) framework. It eliminates the central server by adopting a topology-aware peer-to-peer communication architecture; clients exchange only lightweight head models and locally construct personalized MoE ensembles for multi-expert decision-making. The framework integrates federated knowledge distillation with an adaptive aggregation strategy. This design substantially mitigates knowledge dilution while enhancing model robustness and generalization. Extensive experiments across diverse medical tasks demonstrate consistent superiority over state-of-the-art methods: average accuracy improves by 4.7%, communication overhead decreases by 32%, fault recovery is achieved within milliseconds, and the framework supports both homogeneous and heterogeneous client model configurations.

Technology Category

Application Category

📝 Abstract
Federated learning has wide applications in the medical field. It enables knowledge sharing among different healthcare institutes while protecting patients' privacy. However, existing federated learning systems are typically centralized, requiring clients to upload client-specific knowledge to a central server for aggregation. This centralized approach would integrate the knowledge from each client into a centralized server, and the knowledge would be already undermined during the centralized integration before it reaches back to each client. Besides, the centralized approach also creates a dependency on the central server, which may affect training stability if the server malfunctions or connections are unstable. To address these issues, we propose a decentralized federated learning framework named dFLMoE. In our framework, clients directly exchange lightweight head models with each other. After exchanging, each client treats both local and received head models as individual experts, and utilizes a client-specific Mixture of Experts (MoE) approach to make collective decisions. This design not only reduces the knowledge damage with client-specific aggregations but also removes the dependency on the central server to enhance the robustness of the framework. We validate our framework on multiple medical tasks, demonstrating that our method evidently outperforms state-of-the-art approaches under both model homogeneity and heterogeneity settings.
Problem

Research questions and friction points this paper is trying to address.

Decentralized federated learning for medical data analysis
Reduces knowledge damage in client-specific aggregations
Enhances robustness by removing central server dependency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decentralized federated learning framework
Client-specific Mixture of Experts (MoE)
Direct lightweight head model exchange
🔎 Similar Papers
No similar papers found.
Luyuan Xie
Luyuan Xie
Peking University
AI;Healthcare;Federated Learning
Tianyu Luan
Tianyu Luan
University at Buffalo
3D vision
W
Wenyuan Cai
School of Software and Microelectronics, Peking University; PKU-OCTA Laboratory for Blockchain and Privacy Computing; National Engineering Research Center for Software Engineering, Peking University
Guochen Yan
Guochen Yan
Peking University
GraphFederated learningTrustworthy AI
Zhaoyu Chen
Zhaoyu Chen
TikTok
AI SecurityTrustworthy AIMultimodal AIGenerative AI
Nan Xi
Nan Xi
University at Buffalo
Computer VisionPattern RecognitionMedical AI
Y
Yuejian Fang
School of Software and Microelectronics, Peking University; PKU-OCTA Laboratory for Blockchain and Privacy Computing; National Engineering Research Center for Software Engineering, Peking University
Q
Qingni Shen
School of Software and Microelectronics, Peking University; PKU-OCTA Laboratory for Blockchain and Privacy Computing; National Engineering Research Center for Software Engineering, Peking University
Z
Zhonghai Wu
School of Software and Microelectronics, Peking University; PKU-OCTA Laboratory for Blockchain and Privacy Computing; National Engineering Research Center for Software Engineering, Peking University
Junsong Yuan
Junsong Yuan
State University of New York at Buffalo
computer visionvideo analyticsaction and gesture analysismultimediapattern recognition