FedALT: Federated Fine-Tuning through Adaptive Local Training with Rest-of-the-World LoRA

πŸ“… 2025-03-14
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Federated LoRA fine-tuning based on FedAvg suffers from cross-client interference during model aggregation and struggles to balance personalization with global knowledge acquisition. Method: This paper proposes a novel federated LoRA paradigm that eliminates the need for global initialization. It introduces (1) β€œRest-of-the-World LoRA”, a selective adapter-sharing mechanism enabling low-rank adapter exchange among clients, and (2) a MoE-based adaptive mixer that dynamically gates individual adapter updates and global knowledge integration via learnable weights. Contribution/Results: The method preserves decentralized local training, computational efficiency, and privacy guarantees. Evaluated on NLP benchmarks, it significantly outperforms existing federated LoRA approaches, achieving state-of-the-art performance on local tasks while maintaining robust global model convergence.

Technology Category

Application Category

πŸ“ Abstract
Fine-tuning large language models (LLMs) in federated settings enables privacy-preserving adaptation but suffers from cross-client interference due to model aggregation. Existing federated LoRA fine-tuning methods, primarily based on FedAvg, struggle with data heterogeneity, leading to harmful cross-client interference and suboptimal personalization. In this work, we propose extbf{FedALT}, a novel personalized federated LoRA fine-tuning algorithm that fundamentally departs from FedAvg. Instead of using an aggregated model to initialize local training, each client continues training its individual LoRA while incorporating shared knowledge through a separate Rest-of-the-World (RoTW) LoRA component. To effectively balance local adaptation and global information, FedALT introduces an adaptive mixer that dynamically learns input-specific weightings between the individual and RoTW LoRA components using the Mixture-of-Experts (MoE) principle. Through extensive experiments on NLP benchmarks, we demonstrate that FedALT significantly outperforms state-of-the-art personalized federated LoRA fine-tuning methods, achieving superior local adaptation without sacrificing computational efficiency.
Problem

Research questions and friction points this paper is trying to address.

Addresses cross-client interference in federated LLM fine-tuning
Improves personalization in heterogeneous data environments
Balances local adaptation with global knowledge integration
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive local training with RoTW LoRA
Dynamic weighting using Mixture-of-Experts
Personalized federated LoRA fine-tuning
πŸ”Ž Similar Papers
No similar papers found.