🤖 AI Summary
This work addresses the accuracy degradation of language models caused by inference optimizations—such as quantization, pruning, or erroneous serialization—without access to the original training data. We propose a lightweight, data-free recovery method grounded in synthetic data generation, logit distillation, and selective LoRA fine-tuning: low-rank adapters are learned only at critical layers to approximate the behavior of the high-precision model. To our knowledge, this is the first approach achieving fully data-free, architecture-agnostic accuracy recovery, compatible with diverse attention mechanisms and degradation sources. Evaluated across multiple small language models, our method restores accuracy by 5–17% on average, substantially enhancing post-deployment performance. It achieves strong parameter efficiency—introducing negligible overhead—and maintains practical engineering viability.
📝 Abstract
Inference optimizations such as quantization, pruning, format and datatype conversion, model export, and serialization can lead to functional degradations in language model task performance. While most efforts on performance recovery for deployment focus on robust quantization techniques, we focus on recovering model accuracies from any sources that degrade model weights, such as improper model serialization. In this work, we propose Recover-LoRA, a lightweight and dataset agnostic method to recover accuracy in degraded models. Recover-LoRA uses synthetic data and logit distillation to learn LoRA adapters on selective layers that facilitate aligning the degraded model to its full precision model. We investigate the utility of Recover-LoRA across a diverse set of small language models (SLMs), including models with varying attention architectures, multi-head attention (MHA) and group-query attention (GQA), as well as several evaluation datasets. Our results show that Recover-LoRA recovers model accuracies by 5-17% on MHA and GQA SLMs.