🤖 AI Summary
This work addresses the challenge of low-quality soft labels and poor supervision in existing data-free, one-shot federated learning methods under non-IID data settings, which stem from naive averaging of global teacher models. To overcome this limitation, the authors propose FedBiCross, a novel framework that achieves effective personalization within a single communication round through a three-stage pipeline: client clustering based on output similarity, bi-level cross-cluster optimization, and personalized knowledge distillation. FedBiCross introduces, for the first time, a dynamically weighted cross-cluster knowledge fusion mechanism that mitigates negative transfer and enhances model personalization. Extensive experiments demonstrate that FedBiCross significantly outperforms state-of-the-art methods across four medical imaging datasets and consistently delivers robust performance gains under varying degrees of data heterogeneity.
📝 Abstract
Data-free knowledge distillation-based one-shot federated learning (OSFL) trains a model in a single communication round without sharing raw data, making OSFL attractive for privacy-sensitive medical applications. However, existing methods aggregate predictions from all clients to form a global teacher. Under non-IID data, conflicting predictions cancel out during averaging, yielding near-uniform soft labels that provide weak supervision for distillation. We propose FedBiCross, a personalized OSFL framework with three stages: (1) clustering clients by model output similarity to form coherent sub-ensembles, (2) bi-level cross-cluster optimization that learns adaptive weights to selectively leverage beneficial cross-cluster knowledge while suppressing negative transfer, and (3) personalized distillation for client-specific adaptation. Experiments on four medical image datasets demonstrate that FedBiCross consistently outperforms state-of-the-art baselines across different non-IID degrees.