LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization

📅 2025-07-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high computational overhead, limited expressivity, and catastrophic forgetting inherent in low-rank adaptation (LoRA) and similar methods for domain specialization and continual learning, this paper proposes LoSiA: a dynamic subnet localization and high-rank adaptation framework grounded in gradient sparsity analysis. LoSiA abandons fixed low-rank parameterization and instead identifies sparse, task-critical trainable subnetworks via gradient sparsity patterns, applying high-rank updates exclusively to these subnets. We further introduce LoSiA-Pro, an optimized implementation that reduces training latency. Experiments demonstrate that LoSiA achieves performance on par with full fine-tuning on domain adaptation and commonsense reasoning benchmarks, while attaining the fastest training speed—reducing training latency by approximately 27% compared to LoRA—and significantly mitigating knowledge forgetting during continual learning.

Technology Category

Application Category

📝 Abstract
Parameter-Efficient Fine-Tuning (PEFT) methods, such as LoRA, significantly reduce the number of trainable parameters by introducing low-rank decomposition matrices. However, existing methods perform extensive matrix multiplications in domain specialization tasks, resulting in computational inefficiency and sub-optimal fine-tuning performance. Hence, we propose LoSiA(Low-Resources Subnet Integration Adaptation), an innovative method that dynamically localizes and optimizes critical parameters during the training process. Specifically, it identifies a sub-network using gradient sparsity analysis and optimizes it as the trainable target. This design enables effective high-rank adaptation by updating only the sub-network parameters, reducing the additional matrix multiplication. We also present LoSiA-Pro, a faster implementation of LoSiA, which reduces the training latency by about $27%$ compared to LoRA. Extensive evaluations show that our method achieves minimal performance drop compared to full fine-tuning, while requiring the least training time across domain specialization and common-sense reasoning tasks. Further analysis shows that LoSiA also reduces forgetting during continued training.
Problem

Research questions and friction points this paper is trying to address.

Reduces computational inefficiency in high-rank fine-tuning
Improves fine-tuning performance via dynamic subnet optimization
Minimizes training time while maintaining model accuracy
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic subnet localization via gradient sparsity
Optimizes only critical sub-network parameters
Reduces matrix multiplications for efficiency
🔎 Similar Papers
No similar papers found.
X
Xujia Wang
Tsinghua University, Department of Computer Science and Technology
Yunjia Qi
Yunjia Qi
Tsinghua University
Large Language ModelsInformation ExtractionNatural Language Processing
B
Bin Xu
Tsinghua University