🤖 AI Summary
This work addresses two key limitations of Low-Rank Adaptation (LoRA): suboptimal computational resource allocation and poor robustness of sensitivity metrics. We propose AdaLoRA, the first Bayesian-inspired framework for adaptive LoRA. Its core innovation is a theoretical linkage between parameter sensitivity and the Bayesian signal-to-noise ratio (SNR), proving that parameter magnitude—not variance—dominates importance; thus, SNR replaces conventional sensitivity measures. Integrating the IVON optimizer with a Bayesian pruning mechanism, our method enables more robust and efficient adaptive low-rank updates. Experiments across multiple large language model fine-tuning tasks demonstrate that AdaLoRA matches or exceeds the performance of the original AdaLoRA while training significantly faster than AdaLoRA+Adam. The framework further offers enhanced interpretability, generalizability, and practical applicability.
📝 Abstract
Motivated by the sensitivity-based importance score of the adaptive low-rank adaptation (AdaLoRA), we utilize more theoretically supported metrics, including the signal-to-noise ratio (SNR), along with the Improved Variational Online Newton (IVON) optimizer, for adaptive parameter budget allocation. The resulting Bayesian counterpart not only has matched or surpassed the performance of using the sensitivity-based importance metric but is also a faster alternative to AdaLoRA with Adam. Our theoretical analysis reveals a significant connection between the two metrics, providing a Bayesian perspective on the efficacy of sensitivity as an importance score. Furthermore, our findings suggest that the magnitude, rather than the variance, is the primary indicator of the importance of parameters.