🤖 AI Summary
In federated semi-supervised learning (FSSL), data heterogeneity degrades pseudo-label quality and exacerbates discrepancies between local and global model predictions. To address this, we propose a confidence-difference-driven collaborative training framework that dynamically rectifies pseudo-labels. We first uncover the intrinsic mechanisms by which heterogeneity induces pseudo-label misalignment and divergent prediction biases across clients. Second, we design an adaptive pseudo-label correction strategy guided by inter-client confidence divergence to enhance model consistency. Third, we introduce the Stacked Aggregation with Global Enhancement (SAGE) mechanism to strengthen cross-client knowledge fusion. Evaluated on multiple heterogeneous benchmark datasets, our method achieves 23% faster convergence and a 4.7% average accuracy gain over state-of-the-art FSSL approaches. The implementation is publicly available.
📝 Abstract
Federated Semi-Supervised Learning (FSSL) aims to leverage unlabeled data across clients with limited labeled data to train a global model with strong generalization ability. Most FSSL methods rely on consistency regularization with pseudo-labels, converting predictions from local or global models into hard pseudo-labels as supervisory signals. However, we discover that the quality of pseudo-label is largely deteriorated by data heterogeneity, an intrinsic facet of federated learning. In this paper, we study the problem of FSSL in-depth and show that (1) heterogeneity exacerbates pseudo-label mismatches, further degrading model performance and convergence, and (2) local and global models' predictive tendencies diverge as heterogeneity increases. Motivated by these findings, we propose a simple and effective method called Semi-supervised Aggregation for Globally-Enhanced Ensemble (SAGE), that can flexibly correct pseudo-labels based on confidence discrepancies. This strategy effectively mitigates performance degradation caused by incorrect pseudo-labels and enhances consensus between local and global models. Experimental results demonstrate that SAGE outperforms existing FSSL methods in both performance and convergence. Our code is available at https://github.com/Jay-Codeman/SAGE