FedSRD: Sparsify-Reconstruct-Decompose for Communication-Efficient Federated Large Language Models Fine-Tuning

📅 2025-10-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high communication overhead, structural redundancy of LoRA parameters under network heterogeneity, and aggregation conflicts in federated fine-tuning of large language models (LLMs), this paper proposes the Sparsify-Reconstruct-Decompose (SRD) framework. First, LoRA updates are sparsified via importance-aware pruning; second, sparse updates are reconstructed in the full-rank space prior to aggregation, mitigating heterogeneity-induced conflicts; third, aggregated updates are broadcast efficiently via low-rank decomposition. SRD is the first framework to systematically disentangle LoRA’s structural redundancy, establishing a symmetric communication cycle. A lightweight variant, FedSRD-e, is further introduced to reduce local computational cost. Extensive experiments across ten benchmarks demonstrate up to 90% reduction in communication cost, alongside significantly improved convergence and generalization under heterogeneous data distributions.

Technology Category

Application Category

📝 Abstract
The current paradigm of training large language models (LLMs) on publicly available Web data is becoming unsustainable, with high-quality data sources in specialized domains nearing exhaustion. Federated Learning (FL) emerges as a practical solution for the next generation of AI on a decentralized Web, enabling privacy-preserving collaborative fine-tuning by leveraging private data distributed across a global client base. While Low-Rank Adaptation (LoRA) is the standard for efficient fine-tuning, its application in federated settings presents a critical challenge: communication overhead remains a significant bottleneck across the Web's heterogeneous network conditions. The structural redundancy within LoRA parameters not only incurs a heavy communication burden but also introduces conflicts when aggregating client updates. To address this, we propose FedSRD, a Sparsify-Reconstruct-Decompose framework designed for communication-efficient FL. We first introduce an importance-aware sparsification method that preserves the structural integrity of LoRA updates to reduce the uploaded parameter count. The server then reconstructs and aggregates these updates in a full-rank space to mitigate conflicts. Finally, it decomposes the global update into a sparse low-rank format for broadcast, ensuring a symmetrically efficient cycle. We also propose an efficient variant, FedSRD-e, to reduce computational overhead. Experimental results on 10 benchmarks demonstrate that our framework significantly reduces communication costs by up to 90% while even improving model performance on heterogeneous client data.
Problem

Research questions and friction points this paper is trying to address.

Reduces communication overhead in federated learning for LLMs
Addresses parameter redundancy and update conflicts in LoRA
Enables efficient fine-tuning across heterogeneous network conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Sparsify LoRA updates for reduced communication
Reconstruct and aggregate updates in full-rank space
Decompose global updates into sparse low-rank format
🔎 Similar Papers
No similar papers found.
Guochen Yan
Guochen Yan
Peking University
GraphFederated learningTrustworthy AI
Luyuan Xie
Luyuan Xie
Peking University
AI;Healthcare;Federated Learning
Q
Qingni Shen
Peking University
Y
Yuejian Fang
Peking University
Z
Zhonghai Wu
Peking University