E-3SFC: Communication-Efficient Federated Learning with Double-way Features Synthesizing

📅 2025-02-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the prohibitive communication overhead induced by large models in federated learning, this paper proposes the One-Step Synthetic Feature Compression (3SFC) framework. 3SFC introduces a novel gradient compression paradigm wherein the client model itself serves as the decompressor, integrating inverse gradient reconstruction guided by weight and objective-function priors, one-step synthetic feature generation, error feedback, and dual-path expansion (E-3SFC), enabling dynamic communication budget scheduling. We provide theoretical convergence guarantees for both convex and non-convex settings. Extensive experiments across six benchmark datasets and six model architectures demonstrate that 3SFC outperforms state-of-the-art methods by up to 13.4% in accuracy while reducing total communication volume by up to 111.6×. The framework thus achieves substantial improvements in communication efficiency without compromising model performance.

Technology Category

Application Category

📝 Abstract
The exponential growth in model sizes has significantly increased the communication burden in Federated Learning (FL). Existing methods to alleviate this burden by transmitting compressed gradients often face high compression errors, which slow down the model's convergence. To simultaneously achieve high compression effectiveness and lower compression errors, we study the gradient compression problem from a novel perspective. Specifically, we propose a systematical algorithm termed Extended Single-Step Synthetic Features Compressing (E-3SFC), which consists of three sub-components, i.e., the Single-Step Synthetic Features Compressor (3SFC), a double-way compression algorithm, and a communication budget scheduler. First, we regard the process of gradient computation of a model as decompressing gradients from corresponding inputs, while the inverse process is considered as compressing the gradients. Based on this, we introduce a novel gradient compression method termed 3SFC, which utilizes the model itself as a decompressor, leveraging training priors such as model weights and objective functions. 3SFC compresses raw gradients into tiny synthetic features in a single-step simulation, incorporating error feedback to minimize overall compression errors. To further reduce communication overhead, 3SFC is extended to E-3SFC, allowing double-way compression and dynamic communication budget scheduling. Our theoretical analysis under both strongly convex and non-convex conditions demonstrates that 3SFC achieves linear and sub-linear convergence rates with aggregation noise. Extensive experiments across six datasets and six models reveal that 3SFC outperforms state-of-the-art methods by up to 13.4% while reducing communication costs by 111.6 times. These findings suggest that 3SFC can significantly enhance communication efficiency in FL without compromising model performance.
Problem

Research questions and friction points this paper is trying to address.

Addresses communication burden in Federated Learning
Proposes E-3SFC for efficient gradient compression
Reduces communication costs without performance loss
Innovation

Methods, ideas, or system contributions that make the work stand out.

Double-way gradient compression
Single-step synthetic features
Dynamic communication scheduling
🔎 Similar Papers
No similar papers found.
Y
Yuhao Zhou
College of Computer Science, Sichuan University, Chengdu 610065, P. R. China and Engineering Research Center of Machine Learning and Industry Intelligence, Ministry of Education, Chengdu 610065, P. R. China
Yuxin Tian
Yuxin Tian
Ph.d Candidate, Sichuan University
Deep LearningMachine Learning
Mingjia Shi
Mingjia Shi
Somewhere on the Earth
Learning TheoryData ScienceResource Preserving
Yuanxi Li
Yuanxi Li
University of Illinois Urbana-Champaign
Artificial IntelligenceNatural language processingMachine Learning
Y
Yanan Sun
College of Computer Science, Sichuan University, Chengdu 610065, P. R. China and Engineering Research Center of Machine Learning and Industry Intelligence, Ministry of Education, Chengdu 610065, P. R. China
Qing Ye
Qing Ye
四川大学
Jiancheng Lv
Jiancheng Lv
University of Science and Technology of China
Operations ManagementMarketing