When Does Pruning Benefit Vision Representations?

📅 2025-07-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study systematically investigates the tripartite impact of model pruning on visual representations—specifically interpretability, unsupervised object discovery capability, and alignment with human perception. Method: We conduct experiments across multiple mainstream architectures of varying parameter scales and sparsity levels, integrating feature attribution analysis, representation structure metrics, and human behavioral benchmarking. Contribution/Results: We identify a “sweet spot” in sparsity: moderate (non-extreme) pruning simultaneously enhances attribution interpretability, structural coherence in object discovery, and alignment with human attention patterns—effects modulated by architectural complexity. In contrast, excessive pruning degrades performance. Critically, we reveal for the first time a non-monotonic coupling among sparsity level, architectural scale, and representation quality. These findings provide novel empirical evidence and design principles for interpretable AI and biologically plausible representation learning.

Technology Category

Application Category

📝 Abstract
Pruning is widely used to reduce the complexity of deep learning models, but its effects on interpretability and representation learning remain poorly understood. This paper investigates how pruning influences vision models across three key dimensions: (i) interpretability, (ii) unsupervised object discovery, and (iii) alignment with human perception. We first analyze different vision network architectures to examine how varying sparsity levels affect feature attribution interpretability methods. Additionally, we explore whether pruning promotes more succinct and structured representations, potentially improving unsupervised object discovery by discarding redundant information while preserving essential features. Finally, we assess whether pruning enhances the alignment between model representations and human perception, investigating whether sparser models focus on more discriminative features similarly to humans. Our findings also reveal the presence of sweet spots, where sparse models exhibit higher interpretability, downstream generalization and human alignment. However, these spots highly depend on the network architectures and their size in terms of trainable parameters. Our results suggest a complex interplay between these three dimensions, highlighting the importance of investigating when and how pruning benefits vision representations.
Problem

Research questions and friction points this paper is trying to address.

How pruning affects vision model interpretability and feature attribution
Whether pruning improves unsupervised object discovery via structured representations
If pruning enhances model-human perception alignment by focusing on discriminative features
Innovation

Methods, ideas, or system contributions that make the work stand out.

Analyzing pruning effects on interpretability methods
Exploring pruning for structured unsupervised object discovery
Assessing pruning alignment with human perception
E
Enrico Cassano
University of Turin, Computer Science Department
Riccardo Renzulli
Riccardo Renzulli
PostDoc researcher, University of Turin
Deep Learning
A
Andrea Bragagnolo
University of Turin, Computer Science Department
Marco Grangetto
Marco Grangetto
Full Professor, Università di Torino
Media coding and communicationswireless networks