Automatic Pruning via Structured Lasso with Class-wise Information

📅 2025-02-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing structured pruning methods neglect inter-class statistical disparities, leading to loss of discriminative information. To address this, we propose a semantic-aware structured Lasso pruning framework grounded in the information bottleneck (IB) theory—marking the first integration of class-conditional feature statistics into structured sparsity constraints. Our framework comprises two adaptive variants: graph-guided (sGLP-IB) and tree-guided (sTLP-IB) pruning, both leveraging structural regularization (graph or hierarchical tree) to enforce class-sensitive channel selection and explicitly preserve inter-class separability during compression. On CIFAR-10 with VGG-16, our method achieves 85% parameter reduction, 61% FLOPs reduction, and a 0.14% accuracy gain; on ImageNet with ResNet-50, it reduces parameters by 55% while incurring only a 0.03% top-1 accuracy drop—substantially outperforming state-of-the-art approaches.

Technology Category

Application Category

📝 Abstract
Most pruning methods concentrate on unimportant filters of neural networks. However, they face the loss of statistical information due to a lack of consideration for class-wise data. In this paper, from the perspective of leveraging precise class-wise information for model pruning, we utilize structured lasso with guidance from Information Bottleneck theory. Our approach ensures that statistical information is retained during the pruning process. With these techniques, we introduce two innovative adaptive network pruning schemes: sparse graph-structured lasso pruning with Information Bottleneck ( extbf{sGLP-IB}) and sparse tree-guided lasso pruning with Information Bottleneck ( extbf{sTLP-IB}). The key aspect is pruning model filters using sGLP-IB and sTLP-IB to better capture class-wise relatedness. Compared to multiple state-of-the-art methods, our approaches demonstrate superior performance across three datasets and six model architectures in extensive experiments. For instance, using the VGG16 model on the CIFAR-10 dataset, we achieve a parameter reduction of 85%, a decrease in FLOPs by 61%, and maintain an accuracy of 94.10% (0.14% higher than the original model); we reduce the parameters by 55% with the accuracy at 76.12% using the ResNet architecture on ImageNet (only drops 0.03%). In summary, we successfully reduce model size and computational resource usage while maintaining accuracy. Our codes are at https://anonymous.4open.science/r/IJCAI-8104.
Problem

Research questions and friction points this paper is trying to address.

Improving neural network pruning with class-wise data
Retaining statistical information during pruning process
Introducing adaptive pruning schemes sGLP-IB and sTLP-IB
Innovation

Methods, ideas, or system contributions that make the work stand out.

Utilizes structured lasso for pruning
Integrates Information Bottleneck theory
Develops sGLP-IB and sTLP-IB schemes
🔎 Similar Papers
No similar papers found.
X
Xiang Liu
National University of Singapore
M
Mingchen Li
University of North Texas
X
Xia Li
Department of Information Technology and Electrical Engineering, ETH Zürich
Leigang Qu
Leigang Qu
National University of Singapore
Multimedia ComputingInformation Retrieval
Zifan Peng
Zifan Peng
Ph.D. Candidate at HKUST(GZ)
DeFiTrustworthy AI
Y
Yijun Song
Dongfang College, Zhejiang University of Finance & Economics
Zemin Liu
Zemin Liu
Zhejiang University
Graph LearningGraph Imbalanced Learning
Linshan Jiang
Linshan Jiang
Research Fellow, Institute of Data Science (IDS), NUS
Privacy_preserving_Machine_learningCollaborative Machine LearningEdge-Cloud CollaborationWeb3
J
Jialin Li
National University of Singapore