Optimizing Personalized Federated Learning through Adaptive Layer-Wise Learning

📅 2024-12-10
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address inefficient global knowledge injection and excessive local model personalization—leading to degraded generalization—in personalized federated learning (pFL), this paper proposes FLAYER, the first layer-aware pFL framework. Its core contributions are: (1) layer-adaptive initialization, enabling on-demand injection of global knowledge into specific layers; (2) dynamic layered learning rate scheduling, jointly optimizing personalization and generalization; and (3) selective layered parameter upload and aggregation, enhancing preservation of global representations. Evaluated on four non-IID benchmark datasets spanning computer vision and natural language processing, FLAYER achieves an average inference accuracy improvement of 5.40% (up to +14.29%) over six state-of-the-art methods. It delivers superior personalized modeling performance while maintaining low communication and computational overhead.

Technology Category

Application Category

📝 Abstract
Real-life deployment of federated Learning (FL) often faces non-IID data, which leads to poor accuracy and slow convergence. Personalized FL (pFL) tackles these issues by tailoring local models to individual data sources and using weighted aggregation methods for client-specific learning. However, existing pFL methods often fail to provide each local model with global knowledge on demand while maintaining low computational overhead. Additionally, local models tend to over-personalize their data during the training process, potentially dropping previously acquired global information. We propose FLAYER, a novel layer-wise learning method for pFL that optimizes local model personalization performance. FLAYER considers the different roles and learning abilities of neural network layers of individual local models. It incorporates global information for each local model as needed to initialize the local model cost-effectively. It then dynamically adjusts learning rates for each layer during local training, optimizing the personalized learning process for each local model while preserving global knowledge. Additionally, to enhance global representation in pFL, FLAYER selectively uploads parameters for global aggregation in a layer-wise manner. We evaluate FLAYER on four representative datasets in computer vision and natural language processing domains. Compared to six state-of-the-art pFL methods, FLAYER improves the inference accuracy, on average, by 5.40% (up to 14.29%).
Problem

Research questions and friction points this paper is trying to address.

Optimizing personalized federated learning for non-IID data challenges
Balancing local model personalization with global knowledge retention
Reducing computational overhead in adaptive layer-wise learning methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive layer-wise learning for pFL optimization
Dynamic learning rate adjustment per layer
Selective layer-wise parameter upload for aggregation
🔎 Similar Papers
No similar papers found.
W
Weihang Chen
Shaanxi Normal University, Xian, China
J
Jie Ren
Shaanxi Normal University, Xian, China
Zhiqiang Li
Zhiqiang Li
University of Nebraska-Lincoln
Ling Gao
Ling Gao
Alibaba Group | ShanghaiTech University
Robotic PerceptionEvent-based VisionSLAM
Z
Zheng Wang
University of Leeds, Leeds, United Kingdom