🤖 AI Summary
Traditional parameter-efficient fine-tuning (PEFT) methods—e.g., LoRA—are tightly coupled with specific model architectures, hindering their transfer across heterogeneous large language models (LLMs). This work proposes LoRA-Align and LoRA-Shift, a two-component framework enabling **data-free, training-free** cross-architecture LoRA module transfer. LoRA-Align achieves source-to-target subspace alignment via truncated SVD and Frobenius-optimal linear transformation, while LoRA-Shift adapts the aligned LoRA weights through low-rank projection. The entire process completes in under 20 minutes on consumer-grade GPUs. Evaluated on commonsense reasoning benchmarks—including ARC, OBOA, and HellaSwag—the method yields up to +5.26% absolute improvement over base models, matching the performance of fully fine-tuned LoRA adapters. Crucially, it eliminates the need for task-specific data or retraining, substantially alleviating the generalization bottleneck of PEFT across diverse LLM architectures.
📝 Abstract
Traditional parameter-efficient fine-tuning (PEFT) methods such as LoRA are tightly coupled with the base model architecture, which constrains their applicability across heterogeneous pretrained large language models (LLMs). To address this limitation, we introduce Cross-LoRA, a data-free framework for transferring LoRA modules between diverse base models without requiring additional training data. Cross-LoRA consists of two key components: (a) LoRA-Align, which performs subspace alignment between source and target base models through rank-truncated singular value decomposition (SVD) and Frobenius-optimal linear transformation, ensuring compatibility under dimension mismatch; and (b) LoRA-Shift, which applies the aligned subspaces to project source LoRA weight updates into the target model parameter space. Both components are data-free, training-free, and enable lightweight adaptation on a commodity GPU in 20 minutes. Experiments on ARCs, OBOA and HellaSwag show that Cross-LoRA achieves relative gains of up to 5.26% over base models. Across other commonsense reasoning benchmarks, Cross-LoRA maintains performance comparable to that of directly trained LoRA adapters.