🤖 AI Summary
This work proposes StructLoRA, a novel parameter-efficient fine-tuning method that jointly addresses semantic drift and inter-layer structural inconsistency—two key limitations of Low-Rank Adaptation (LoRA). StructLoRA introduces an information bottleneck-guided filter to suppress task-irrelevant directions and a lightweight graph coordinator to align layer-wise updates during training. The approach incurs no additional inference overhead and achieves state-of-the-art performance across diverse architectures, including LLaMA, LLaVA, and ViT. Notably, StructLoRA demonstrates substantial improvements over existing methods under low-rank and few-shot settings, where maintaining semantic fidelity and structural coherence is particularly challenging.
📝 Abstract
Low-Rank Adaptation (LoRA) has become a cornerstone of parameter-efficient fine-tuning (PEFT). Yet, its efficacy is hampered by two fundamental limitations: semantic drift, by treating all update directions with equal importance, and structural incoherence, from adapting layers independently, resulting in suboptimal, uncoordinated updates. To remedy these, we propose StructLoRA, a framework that addresses both limitations through a principled, dual-component design: (1) an Information Bottleneck-guided filter that prunes task-irrelevant directions to mitigate semantic drift, and (2) a lightweight, training-only graph-based coordinator that enforces inter-layer consistency to resolve structural incoherence. Extensive experiments across large language model , vision language model, and vision model (including LLaMA, LLaVA, and ViT) demonstrate that StructLoRA consistently establishes a new state-of-the-art, outperforming not only vanilla LoRA but also advanced dynamic rank allocation and sparsity-based methods. Notably, the benefits are particularly pronounced in challenging low-rank and low-data regimes. Crucially, since our proposed modules operate only during training, StructLoRA enhances performance with zero additional inference cost, advancing the focus of PEFT -- from mere parameter compression to a more holistic optimization of information quality and structural integrity.