Near-Optimal Decentralized Stochastic Nonconvex Optimization with Heavy-Tailed Noise

πŸ“… 2026-01-16
πŸ“ˆ Citations: 2
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses decentralized non-convex optimization over row-stochastic networks in the presence of heavy-tailed gradient noise. The authors propose a novel algorithm that integrates normalized stochastic gradient descent with Pull-Diag gradient tracking. To the best of our knowledge, this is the first method to simultaneously achieve optimal sample complexity and near-optimal communication complexity under heavy-tailed noise, applicable to both directed and undirected network topologies. Theoretical analysis establishes nearly tight upper bounds on these complexities, while empirical evaluations demonstrate the algorithm’s superior performance in practical scenarios.

Technology Category

Application Category

πŸ“ Abstract
This paper studies decentralized stochastic nonconvex optimization problem over row-stochastic networks. We consider the heavy-tailed gradient noise which is empirically observed in many popular real-world applications. Specifically, we propose a decentralized normalized stochastic gradient descent with Pull-Diag gradient tracking, which achieves approximate stationary points with the optimal sample complexity and the near-optimal communication complexity. We further follow our framework to study the setting of undirected networks, also achieving the nearly tight upper complexity bounds. Moreover, we conduct empirical studies to show the practical superiority of the proposed methods.
Problem

Research questions and friction points this paper is trying to address.

decentralized optimization
stochastic nonconvex optimization
heavy-tailed noise
row-stochastic networks
Innovation

Methods, ideas, or system contributions that make the work stand out.

decentralized optimization
heavy-tailed noise
normalized SGD
gradient tracking
nonconvex optimization
πŸ”Ž Similar Papers
No similar papers found.