Compressed Proximal Federated Learning for Non-Convex Composite Optimization on Heterogeneous Data

📅 2026-03-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes FedCEF, an algorithm designed to address key challenges in federated composite optimization under heterogeneous data, including low communication efficiency, difficulty in handling non-smooth regularizers, and convergence instability caused by compression bias. FedCEF decouples proximal updates from communication, enabling clients to handle non-smooth terms locally, and introduces a downlink strategy that allows exact reconstruction of the global control variable without explicit transmission. Furthermore, it incorporates an error feedback mechanism with control variates to effectively mitigate the adverse effects of quantization noise and data heterogeneity. Theoretical analysis establishes that FedCEF converges to a controllable residual error under non-convex settings, and experiments demonstrate its ability to maintain high accuracy even at an extreme 1% compression rate, significantly reducing communication overhead.

Technology Category

Application Category

📝 Abstract
Federated Composite Optimization (FCO) has emerged as a promising framework for training models with structural constraints (e.g., sparsity) in distributed edge networks. However, simultaneously achieving communication efficiency and convergence robustness remains a significant challenge, particularly when dealing with non-smooth regularizers, statistical heterogeneity, and the restrictions of biased compression. To address these issues, we propose FedCEF (Federated Composite Error Feedback), a novel algorithm tailored for non-convex FCO. FedCEF introduces a decoupled proximal update scheme that separates the proximal operator from communication, enabling clients to handle non-smooth terms locally while transmitting compressed information. To mitigate the noise from aggressive quantization and the bias from non-IID data, FedCEF integrates a rigorous error feedback mechanism with control variates. Furthermore, we design a communication-efficient pre-proximal downlink strategy that allows clients to exactly reconstruct global control variables without explicit transmission. We theoretically establish that FedCEF achieves sublinear convergence to a bounded residual error under general non-convexity, which is controllable via the step size and batch size. Extensive experiments on real datasets validate FedCEF maintains competitive model accuracy even under extreme compression ratios (e.g., 1%), significantly reducing the total communication volume compared to uncompressed baselines.
Problem

Research questions and friction points this paper is trying to address.

Federated Composite Optimization
Non-Convex Optimization
Communication Efficiency
Statistical Heterogeneity
Biased Compression
Innovation

Methods, ideas, or system contributions that make the work stand out.

Federated Composite Optimization
Error Feedback
Proximal Decoupling
Communication Compression
Non-Convex Optimization
🔎 Similar Papers
No similar papers found.
P
Pu Qiu
School of Intelligent Systems Engineering, Sun Yat-Sen University, Shenzhen 518107, P.R China
C
Chen Ouyang
School of Intelligent Systems Engineering, Sun Yat-Sen University, Shenzhen 518107, P.R China
Y
Yongyang Xiong
School of Intelligent Systems Engineering, Sun Yat-Sen University, Shenzhen 518107, P.R China
K
Keyou You
Department of Automation, Beijing National Research Center for Information Science and Technology, Tsinghua University, Beijing 100084, China
Wanquan Liu
Wanquan Liu
Sun Yat-sen University
Computer visionIntelligent controlPattern recognition
Y
Yang Shi
Department of Mechanical Engineering, University of Victoria, Victoria, BC V8W 2Y2, Canada