π€ AI Summary
This work investigates the theoretical relationship between the update rank in LoRA (Low-Rank Adaptation) fine-tuning and model accuracy, elucidating why high performance can be maintained even at extremely low ranksβsuch as rank-1. Adopting a dynamical systems perspective, the study rigorously derives the gradient flow equations for LoRA for the first time, proving the equivalence between synchronous and sequential update schemes. Furthermore, under both trace-square loss and Frobenius norm loss, it establishes explicit closed-form relationships between rank and approximation accuracy. These results provide a principled theoretical foundation for rank selection in LoRA, quantifying the achievable accuracy at different ranks and demonstrating that low-rank updates are theoretically sufficient to approximate the performance of full-parameter fine-tuning.
π Abstract
Previous empirical studies have shown that LoRA achieves accuracy comparable to full-parameter methods on downstream fine-tuning tasks, even for rank-1 updates. By contrast, the theoretical underpinnings of the dependence of LoRA's accuracy on update rank remain relatively unexplored. In this work, we compare the accuracy of rank-r LoRA updates against full-parameter updates for fine-tuning tasks from a dynamical systems perspective. We perform gradient flow analysis in both full-rank and low-rank regimes to establish explicit relationships between rank and accuracy for two loss functions under LoRA. While gradient flow equations for LoRA are presented in prior work, we rigorously derive their form and show that they are identical for simultaneous and sequential LoRA parameter updates. We then use the resulting dynamical system equations to obtain closed-form relationships between LoRA rank and accuracy for trace-squared and Frobenius-norm low-rank approximation loss functions.