🤖 AI Summary
To address client drift caused by data heterogeneity in federated learning, as well as high communication, computation, and memory overheads, this paper proposes FedSub. The method introduces low-dimensional subspace projection during local updates, constraining model updates to a shared low-dimensional subspace, and employs low-dimensional dual variables to jointly enforce global consistency. Theoretically, we analyze how step size and projection matrices affect convergence rate and provide rigorous convergence guarantees. Empirically, FedSub achieves comparable model accuracy to baseline methods on image and text benchmarks, while reducing communication volume by approximately 60%, decreasing training memory usage by 55%, and significantly lowering computational cost. The algorithm thus offers strong efficiency, provable convergence, and scalability—making it particularly suitable for resource-constrained federated environments.
📝 Abstract
This work addresses the key challenges of applying federated learning to large-scale deep neural networks, particularly the issue of client drift due to data heterogeneity across clients and the high costs of communication, computation, and memory. We propose FedSub, an efficient subspace algorithm for federated learning on heterogeneous data. Specifically, FedSub utilizes subspace projection to guarantee local updates of each client within low-dimensional subspaces, thereby reducing communication, computation, and memory costs. Additionally, it incorporates low-dimensional dual variables to mitigate client drift. We provide convergence analysis that reveals the impact of key factors such as step size and subspace projection matrices on convergence. Experimental results demonstrate its efficiency.