🤖 AI Summary
This work investigates the convergence dynamics of distributed SGD, compressed SGD, and SignSGD under $(L_0, L_1)$-smooth objectives and generalized (including heavy-tailed) gradient noise. Adopting a stochastic differential equation (SDE) framework for continuous-time modeling, we systematically characterize the coupled effects of batch noise, gradient compression, and learning-rate adaptivity. We establish, for the first time, that adaptive methods—including Distributed SignSGD—remain convergent under heavy-tailed noise, whereas standard non-adaptive step-size decay schemes provably fail unless they implicitly depend on gradient norms; this unifies and rigorously justifies the theoretical necessity of adaptivity. Furthermore, we quantify how compression interacts with noise to affect convergence rates, and validate the high fidelity of our SDE approximation via dynamical simulations. The analysis provides a novel theoretical framework and design principles for robust distributed training.
📝 Abstract
Using stochastic differential equation (SDE) approximations, we study the dynamics of Distributed SGD, Distributed Compressed SGD, and Distributed SignSGD under $(L_0,L_1)$-smoothness and flexible noise assumptions. Our analysis provides insights -- which we validate through simulation -- into the intricate interactions between batch noise, stochastic gradient compression, and adaptivity in this modern theoretical setup. For instance, we show that extit{adaptive} methods such as Distributed SignSGD can successfully converge under standard assumptions on the learning rate scheduler, even under heavy-tailed noise. On the contrary, Distributed (Compressed) SGD with pre-scheduled decaying learning rate fails to achieve convergence, unless such a schedule also accounts for an inverse dependency on the gradient norm -- de facto falling back into an adaptive method.