PM-MOE: Mixture of Experts on Private Model Parameters for Personalized Federated Learning

πŸ“… 2025-02-01
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address poor global model generalization and inefficient collaborative utilization of locally private parameters in personalized federated learning (pFL) under statistical heterogeneity, this paper proposes an energy-function-driven Mixture-of-Experts (MoE) framework. Methodologically, it introduces Energy-Based Models (EBMs) into the MoE architecture for the first time, designs a lightweight denoising mechanism to enable trust-aware selection and zero-shot reuse of client-specific modules across devices, and integrates model-splitting-based pFL with cross-device private-parameter distillation. Evaluated on six benchmark datasets under two statistical heterogeneity settings, the method consistently improves the performance of nine state-of-the-art pFL algorithms, achieving average accuracy gains of 1.2–3.8%. Crucially, these improvements incur only negligible communication and computational overhead.

Technology Category

Application Category

πŸ“ Abstract
Federated learning (FL) has gained widespread attention for its privacy-preserving and collaborative learning capabilities. Due to significant statistical heterogeneity, traditional FL struggles to generalize a shared model across diverse data domains. Personalized federated learning addresses this issue by dividing the model into a globally shared part and a locally private part, with the local model correcting representation biases introduced by the global model. Nevertheless, locally converged parameters more accurately capture domain-specific knowledge, and current methods overlook the potential benefits of these parameters. To address these limitations, we propose PM-MoE architecture. This architecture integrates a mixture of personalized modules and an energy-based personalized modules denoising, enabling each client to select beneficial personalized parameters from other clients. We applied the PM-MoE architecture to nine recent model-split-based personalized federated learning algorithms, achieving performance improvements with minimal additional training. Extensive experiments on six widely adopted datasets and two heterogeneity settings validate the effectiveness of our approach. The source code is available at url{https://github.com/dannis97500/PM-MOE}.
Problem

Research questions and friction points this paper is trying to address.

Personalized Federated Learning
Model Performance Enhancement
Device-specific Expertise Utilization
Innovation

Methods, ideas, or system contributions that make the work stand out.

PM-MoE Architecture
Personalized Federated Learning
Expert Models Integration
πŸ”Ž Similar Papers
No similar papers found.
Y
Yu Feng
Beijing Univ. of Posts and Telecomm., Beijing, China
Yangli-ao Geng
Yangli-ao Geng
Beijing Jiaotong University
Machine LearningData MiningUnsupervised Learning
Yifan Zhu
Yifan Zhu
Beijing University of Posts and Telecommunications
PEFT of LLMsGraph RAGGraph mining
Z
Zongfu Han
Beijing Univ. of Posts and Telecomm., Beijing, China
X
Xie Yu
Beijing Univ. of Aeronautics and Astronautics, Beijing, China
K
Kaiwen Xue
Beijing Univ. of Posts and Telecomm., Beijing, China
Haoran Luo
Haoran Luo
Nanyang Technological University
Knowledge GraphLarge Language ModelsGraph Neural Networks
Mengyang Sun
Mengyang Sun
Northwestern Polytechnical University
computer vision、 vision-language interaction
G
Guangwei Zhang
Beijing Univ. of Posts and Telecomm., Beijing, China
Meina Song
Meina Song
Professor of Computer Science, Beijing University of Posts and Telecommunications
data science