🤖 AI Summary
Existing structural pruning methods often neglect topological dependencies among parameters, leading to substantial performance degradation after compression. To address this, we propose a structured pruning framework that jointly optimizes connection importance and network performance. First, we introduce a Jacobian-based criterion to quantify the joint significance of parameters both within and across layers, explicitly modeling parameter interconnectivity. Second, we design an equivalent pruning mechanism that employs a lightweight autoencoder to preserve the gradient contributions of pruned connections, thereby mitigating accuracy loss during fine-tuning. Our approach integrates first-order gradient sensitivity analysis, structured pruning, and reconstruction-aware fine-tuning. Experiments demonstrate that the proposed Jacobian criterion consistently outperforms mainstream importance metrics across multiple benchmarks. Moreover, equivalent pruning yields an average Top-1 accuracy improvement of 2.3%, significantly reducing the performance drop typically incurred between pruning and fine-tuning.
📝 Abstract
Structural pruning has been widely studied for its effectiveness in compressing neural networks. However, existing methods often neglect the interconnections among parameters. To address this limitation, this paper proposes a structural pruning framework termed Optimal Brain Connection. First, we introduce the Jacobian Criterion, a first-order metric for evaluating the saliency of structural parameters. Unlike existing first-order methods that assess parameters in isolation, our criterion explicitly captures both intra-component interactions and inter-layer dependencies. Second, we propose the Equivalent Pruning mechanism, which utilizes autoencoders to retain the contributions of all original connection--including pruned ones--during fine-tuning. Experimental results demonstrate that the Jacobian Criterion outperforms several popular metrics in preserving model performance, while the Equivalent Pruning mechanism effectively mitigates performance degradation after fine-tuning. Code: https://github.com/ShaowuChen/Optimal_Brain_Connection