🤖 AI Summary
To address the challenges in large language model (LLM) pruning—namely, intractable layer-wise mask optimization, scarcity of calibration data, and prohibitively large combinatorial search spaces rendering integer programming (IP) infeasible—this paper proposes an efficient and scalable 1-swap mask fine-tuning method. Our key contributions are threefold: (i) we derive, for the first time, a closed-form optimal solution for 1-swap updates based on the Gram matrix; (ii) we decouple row-wise constraints to enable GPU-native parallelism and eliminate hyperparameter tuning, facilitating large-scale Transformer adaptation; and (iii) we integrate uniform sparsity constraints with IP-inspired dimensionality reduction to avoid full retraining. Evaluated on GPT-family models, our method reduces per-layer pruning error by up to 60% compared to Wanda, while significantly improving perplexity and zero-shot accuracy.
📝 Abstract
The resource requirements of Neural Networks can be significantly reduced through pruning -- the removal of seemingly less important parameters. However, with the rise of Large Language Models (LLMs), full retraining to recover pruning-induced performance degradation is often prohibitive and classical approaches such as global magnitude pruning are suboptimal on Transformer architectures. State-of-the-art methods hence solve a layer-wise mask selection problem, the problem of finding a pruning mask which minimizes the per-layer pruning error on a small set of calibration data. Exactly solving this problem to optimality using Integer Programming (IP) solvers is computationally infeasible due to its combinatorial nature and the size of the search space, and existing approaches therefore rely on approximations or heuristics. In this work, we demonstrate that the mask selection problem can be made drastically more tractable at LLM scale. To that end, we decouple the rows by enforcing equal sparsity levels per row. This allows us to derive optimal 1-swaps (exchanging one kept and one pruned weight) that can be computed efficiently using the Gram matrix of the calibration data. Using these observations, we propose a tractable and simple 1-swap algorithm that warm starts from any pruning mask, runs efficiently on GPUs at LLM scale, and is essentially hyperparameter-free. We demonstrate that our approach reduces per-layer pruning error by up to 60% over Wanda (Sun et al., 2023) and consistently improves perplexity and zero-shot accuracy across state-of-the-art GPT architectures.