🤖 AI Summary
Traditional clustered federated learning (CFL) suffers from fragmented learning and limited inter-cluster knowledge sharing due to isolated model training within clusters. To address this, this paper proposes a hierarchical clustered federated learning framework tailored for IoT environments. The framework introduces a two-tier aggregation mechanism—local clustering modeling at the edge and unified global modeling in the cloud—and pioneers a hierarchical knowledge distillation method that enables multi-teacher, cross-cluster knowledge transfer, thereby preserving cluster-level personalization while integrating collective learning insights. Experiments on standard benchmark datasets demonstrate that the proposed approach significantly outperforms mainstream baselines in both cluster-level and global model accuracy, with improvements of 3.32%–7.57%. This achieves effective co-optimization of personalization capability and generalization performance.
📝 Abstract
Clustered Federated Learning (CFL) has emerged as a powerful approach for addressing data heterogeneity and ensuring privacy in large distributed IoT environments. By clustering clients and training cluster-specific models, CFL enables personalized models tailored to groups of heterogeneous clients. However, conventional CFL approaches suffer from fragmented learning for training independent global models for each cluster and fail to take advantage of collective cluster insights. This paper advocates a shift to hierarchical CFL, allowing bi-level aggregation to train cluster-specific models at the edge and a unified global model at the cloud. This shift improves training efficiency yet might introduce communication challenges. To this end, we propose CFLHKD, a novel personalization scheme for integrating hierarchical cluster knowledge into CFL. Built upon multi-teacher knowledge distillation, CFLHKD enables inter-cluster knowledge sharing while preserving cluster-specific personalization. CFLHKD adopts a bi-level aggregation to bridge the gap between local and global learning. Extensive evaluations of standard benchmark datasets demonstrate that CFLHKD outperforms representative baselines in cluster-specific and global model accuracy and achieves a performance improvement of 3.32-7.57%.