🤖 AI Summary
Non-IID data across clients in federated learning impedes the global model’s ability to achieve satisfactory personalized performance. To address this, we propose a distributed client clustering and collaborative training framework for personalized federated learning. Our core contribution is a lightweight “Lazy Influence” influence function approximation, enabling decentralized, label-free semantic similarity–based client clustering without access to centralized data or labels. Coupled with clustering-aware personalized model aggregation and local fine-tuning, the framework achieves efficient non-IID optimization. Evaluated on Nordic language prediction and CIFAR-100, our method fully recovers the global model’s performance loss under non-IID conditions, matching the performance of oracle clustering and outperforming mainstream baselines by an average of 17%.
📝 Abstract
In Federated Learning, heterogeneity in client data distributions often means that a single global model does not have the best performance for individual clients. Consider for example training a next-word prediction model for keyboards: user-specific language patterns due to demographics (dialect, age, etc.), language proficiency, and writing style result in a highly non-IID dataset across clients. Other examples are medical images taken with different machines, or driving data from different vehicle types. To address this, we propose a simple yet effective personalized federated learning framework (pFedLIA) that utilizes a computationally efficient influence approximation, called `Lazy Influence', to cluster clients in a distributed manner before model aggregation. Within each cluster, data owners collaborate to jointly train a model that captures the specific data patterns of the clients. Our method has been shown to successfully recover the global model's performance drop due to the non-IID-ness in various synthetic and real-world settings, specifically a next-word prediction task on the Nordic languages as well as several benchmark tasks. It matches the performance of a hypothetical Oracle clustering, and significantly improves on existing baselines, e.g., an improvement of 17% on CIFAR100.