🤖 AI Summary
Federated fine-tuning of large language models (LLMs) suffers from high computational and communication overhead, hindering deployment on resource-constrained devices. To address this, we propose SmartFed—a highly efficient federated fine-tuning framework. Its core innovations are: (1) reusing pretrained LoRA modules to avoid training from scratch; (2) introducing a rank-level Mixture-of-Rank-Experts (MoRE) mechanism for semantic-aware, fine-grained knowledge selection; and (3) designing an Elastic Expert Quota Allocation (EEQA) algorithm that dynamically optimizes parameter activation and computational load based on input semantics and device resource budgets. SmartFed integrates LoRA decomposition, rank-level expert activation, gating-based combination, and federated collaborative optimization. Extensive experiments across multiple benchmarks demonstrate that SmartFed significantly improves model performance and training efficiency—reducing communication volume by up to 42% and computational cost by 37%.
📝 Abstract
Federated fine-tuning offers a promising solution for adapting Large Language Models (LLMs) to downstream tasks while safeguarding data privacy. However, its high computational and communication demands hinder its deployment on resource-constrained devices. In this paper, we propose SmartFed, a resource-efficient federated fine-tuning framework. SmartFed intelligently reuses knowledge embedded in existing LoRA modules, eliminating the need for expensive training from scratch when adapting LLMs to new tasks. To effectively exploit this knowledge and ensure scalability, we introduce the Mixture of Rank-Wise Experts (MoRE). MoRE decomposes LoRA modules into fine-grained rank-level experts. These experts are selectively activated and combined based on input semantics and resource budgets. Moreover, to optimize resource utilization, we present the Elastic Expert Quota Allocation (EEQA). EEQA adaptively allocates expert capacity across parameter matrices based on their contribution to model performance, focusing computing resources on the critical experts. Extensive evaluations across multiple benchmarks demonstrate that SmartFed significantly outperforms existing methods in model performance and training efficiency.