π€ AI Summary
In federated learning under high data heterogeneity, severe client drift occurs, and conventional reference-model-based gradient correction methods suffer from instability in partial participation settings.
Method: This paper proposes a reference-model-free hierarchical gradient centralization mechanism. It performs coordinated gradient correction on the feature extraction and classifier layers separately during local training and global aggregation, establishing a local-global two-stage hybrid optimization paradigm.
Contribution/Results: We provide theoretical convergence guarantees showing strict improvement over standard FedAvg. Empirical evaluation under strong heterogeneity demonstrates up to a 20% absolute improvement in test accuracy over baseline methods, significantly outperforming existing approaches and achieving state-of-the-art performance.
π Abstract
Multi-source information fusion (MSIF) leverages diverse data streams to enhance decision-making, situational awareness, and system resilience. Federated Learning (FL) enables MSIF while preserving privacy but suffers from client drift under high data heterogeneity, leading to performance degradation. Traditional mitigation strategies rely on reference-based gradient adjustments, which can be unstable in partial participation settings. To address this, we propose Gradient Centralized Federated Learning (GC-Fed), a reference-free gradient correction method inspired by Gradient Centralization (GC). We introduce Local GC and Global GC, applying GC during local training and global aggregation, respectively. Our hybrid GC-Fed approach selectively applies GC at the feature extraction layer locally and at the classifier layer globally, improving training stability and model performance. Theoretical analysis and empirical results demonstrate that GC-Fed mitigates client drift and achieves state-of-the-art accuracy gains of up to 20% in heterogeneous settings.