🤖 AI Summary
In federated learning, large-scale model training suffers from high communication overhead and difficulty in simultaneously achieving fast convergence and high accuracy. To address this, we propose ProjFL and its enhanced variant ProjFL+EF. Our core innovation is constructing a low-dimensional shared subspace from historical global descent directions and projecting local gradients onto this subspace before compression and transmission—significantly reducing communication cost. ProjFL supports both unbiased and biased compressors, while ProjFL+EF further incorporates error feedback to ensure convergence. We provide rigorous theoretical guarantees for convergence under strongly convex, convex, and non-convex objectives. Extensive experiments on standard image classification benchmarks demonstrate that our methods reduce communication costs by up to 90% while maintaining test accuracy comparable to state-of-the-art baselines.
📝 Abstract
Federated Learning (FL) enables decentralized model training across multiple clients while optionally preserving data privacy. However, communication efficiency remains a critical bottleneck, particularly for large-scale models. In this work, we introduce two complementary algorithms: ProjFL, designed for unbiased compressors, and ProjFL+EF, tailored for biased compressors through an Error Feedback mechanism. Both methods rely on projecting local gradients onto a shared client-server subspace spanned by historical descent directions, enabling efficient information exchange with minimal communication overhead. We establish convergence guarantees for both algorithms under strongly convex, convex, and non-convex settings. Empirical evaluations on standard FL classification benchmarks with deep neural networks show that ProjFL and ProjFL+EF achieve accuracy comparable to existing baselines while substantially reducing communication costs.