Communication-Efficient Federated Learning by Exploiting Spatio-Temporal Correlations of Gradients

📅 2026-01-15
🏛️ IEEE transactions on computers
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the high uplink communication overhead of federated learning in bandwidth-constrained settings, a challenge exacerbated by the common oversight of temporal correlations in gradients across training rounds. To tackle this, we propose GradESTC, a novel gradient compression method that jointly models the spatial low-rank structure of gradients and their temporal correlations over successive rounds. By transmitting only a small set of dynamically updated basis vectors along with lightweight combination coefficients, GradESTC achieves substantial communication savings. Our approach maintains convergence speed and final model accuracy while reducing uplink communication costs by 39.79% on average compared to the strongest baseline, demonstrating significantly improved communication efficiency.

Technology Category

Application Category

📝 Abstract
Communication overhead is a critical challenge in federated learning, particularly in bandwidth-constrained networks. Although many methods have been proposed to reduce communication overhead, most focus solely on compressing individual gradients, overlooking the temporal correlations among them. Prior studies have shown that gradients exhibit spatial correlations, typically reflected in low-rank structures. Through empirical analysis, we further observe a strong temporal correlation between client gradients across adjacent rounds. Based on these observations, we propose GradESTC, a compression technique that exploits both spatial and temporal gradient correlations. GradESTC exploits spatial correlations to decompose each full gradient into a compact set of basis vectors and corresponding combination coefficients. By exploiting temporal correlations, only a small portion of the basis vectors need to be dynamically updated in each round. GradESTC significantly reduces communication overhead by transmitting lightweight combination coefficients and a limited number of updated basis vectors instead of the full gradients. Extensive experiments show that, upon reaching a target accuracy level near convergence, GradESTC reduces uplink communication by an average of 39.79% compared to the strongest baseline, while maintaining comparable convergence speed and final accuracy to uncompressed FedAvg. By effectively leveraging spatio-temporal gradient structures, GradESTC offers a practical and scalable solution for communication-efficient federated learning.
Problem

Research questions and friction points this paper is trying to address.

communication overhead
federated learning
gradient compression
temporal correlation
spatial correlation
Innovation

Methods, ideas, or system contributions that make the work stand out.

federated learning
gradient compression
spatio-temporal correlation
communication efficiency
low-rank decomposition
🔎 Similar Papers
No similar papers found.
S
Shenlong Zheng
College of Information Science and Technology, Jinan University, Guangzhou, China, 510632
Zhen Zhang
Zhen Zhang
University of Chinese Academy of Sciences
Deep LearningSignal Processing
Yuhui Deng
Yuhui Deng
Professor of Computer Science, Jinan University
Cloud ComputingInformation StorageData ManagementComputer System
Geyong Min
Geyong Min
University of Exeter
L
Lin Cui
College of Information Science and Technology, Jinan University, Guangzhou, China, 510632