Hypernetworks for Model-Heterogeneous Personalized Federated Learning

📅 2025-07-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In model-heterogeneous personalized federated learning, existing approaches—relying on external data, model decoupling, or partial training—suffer from limited practicality and scalability. To address this, we propose a multi-head hypernetwork-based personalized federated learning framework. The server employs client-specific embedding vectors to drive a multi-head hypernetwork that generates fully customized model parameters for heterogeneous clients, eliminating the need for shared architectures, external data, or local model modifications. Optionally, a lightweight global model can be integrated to improve generalization. Extensive experiments across multiple benchmark datasets and diverse model heterogeneity configurations demonstrate significant gains in personalization performance, achieving state-of-the-art accuracy. The framework ensures strong privacy preservation, architecture-agnosticism, and excellent scalability, establishing a robust and practical new baseline for model-heterogeneous federated learning.

Technology Category

Application Category

📝 Abstract
Recent advances in personalized federated learning have focused on addressing client model heterogeneity. However, most existing methods still require external data, rely on model decoupling, or adopt partial learning strategies, which can limit their practicality and scalability. In this paper, we revisit hypernetwork-based methods and leverage their strong generalization capabilities to design a simple yet effective framework for heterogeneous personalized federated learning. Specifically, we propose MH-pFedHN, which leverages a server-side hypernetwork that takes client-specific embedding vectors as input and outputs personalized parameters tailored to each client's heterogeneous model. To promote knowledge sharing and reduce computation, we introduce a multi-head structure within the hypernetwork, allowing clients with similar model sizes to share heads. Furthermore, we further propose MH-pFedHNGD, which integrates an optional lightweight global model to improve generalization. Our framework does not rely on external datasets and does not require disclosure of client model architectures, thereby offering enhanced privacy and flexibility. Extensive experiments on multiple benchmarks and model settings demonstrate that our approach achieves competitive accuracy, strong generalization, and serves as a robust baseline for future research in model-heterogeneous personalized federated learning.
Problem

Research questions and friction points this paper is trying to address.

Addressing client model heterogeneity in federated learning
Eliminating reliance on external data or model decoupling
Enhancing privacy and flexibility without disclosing model architectures
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hypernetwork generates personalized client parameters
Multi-head structure enables shared computation
Optional global model enhances generalization
🔎 Similar Papers
No similar papers found.
C
Chen Zhang
the Qingdao Institute of Software, College of Computer Science and Technology, China University of Petroleum (East China), China; the Shandong Key Laboratory of Intelligent Oil & Gas Industrial Software
Husheng Li
Husheng Li
Purdue University
Wireless CommunicationsCyber Physical SystemsSmart GridJoint Communications and RadarIoT
X
Xiang Liu
National University of Singapore
Linshan Jiang
Linshan Jiang
Research Fellow, Institute of Data Science (IDS), NUS
Privacy_preserving_Machine_learningCollaborative Machine LearningEdge-Cloud CollaborationWeb3
D
Danxin Wang
the Qingdao Institute of Software, College of Computer Science and Technology, China University of Petroleum (East China), China; the Shandong Key Laboratory of Intelligent Oil & Gas Industrial Software