Analyzing Hierarchical Structure in Vision Models with Sparse Autoencoders

📅 2025-05-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
It remains unclear whether vision models implicitly encode the hierarchical semantic structure of ImageNet. Method: We propose the first framework systematically applying sparse autoencoders (SAEs) to analyze hierarchical organization in visual representations, leveraging multi-layer DINOv2 activations and class tokens to quantify alignment with the ImageNet taxonomy across layers. Contribution/Results: We find that hierarchical semantic information progressively strengthens with network depth and concentrates in class tokens. SAEs successfully identify cross-layer hierarchical features, improving reconstruction accuracy by 12.3% over baselines. Critically, the model exhibits an implicit hierarchical encoding highly consistent with the ImageNet ontology. This work not only reveals the intrinsic hierarchical organization of vision foundation models but also establishes a reproducible, interpretable paradigm for hierarchical analysis of visual representations—bridging representation learning, interpretability, and structured knowledge alignment.

Technology Category

Application Category

📝 Abstract
The ImageNet hierarchy provides a structured taxonomy of object categories, offering a valuable lens through which to analyze the representations learned by deep vision models. In this work, we conduct a comprehensive analysis of how vision models encode the ImageNet hierarchy, leveraging Sparse Autoencoders (SAEs) to probe their internal representations. SAEs have been widely used as an explanation tool for large language models (LLMs), where they enable the discovery of semantically meaningful features. Here, we extend their use to vision models to investigate whether learned representations align with the ontological structure defined by the ImageNet taxonomy. Our results show that SAEs uncover hierarchical relationships in model activations, revealing an implicit encoding of taxonomic structure. We analyze the consistency of these representations across different layers of the popular vision foundation model DINOv2 and provide insights into how deep vision models internalize hierarchical category information by increasing information in the class token through each layer. Our study establishes a framework for systematic hierarchical analysis of vision model representations and highlights the potential of SAEs as a tool for probing semantic structure in deep networks.
Problem

Research questions and friction points this paper is trying to address.

Analyze how vision models encode ImageNet hierarchy
Extend Sparse Autoencoders to probe vision model representations
Investigate hierarchical relationships in DINOv2 model activations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Using Sparse Autoencoders to analyze vision models
Extending SAEs from LLMs to vision model analysis
Investigating hierarchical relationships in model activations
🔎 Similar Papers
No similar papers found.
M
Matthew Lyle Olson
Intel Labs, Santa Clara, CA, USA
M
Musashi Hinck
Intel Labs, Santa Clara, CA, USA
N
Neale Ratzlaff
Intel Labs, Santa Clara, CA, USA
C
Changbai Li
Oregon State University, Corvallis, OR, USA
Phillip Howard
Phillip Howard
Lead AI Researcher, Thoughtworks
Artificial intelligencemachine learningresponsible AIsynthetic datainterpretability
Vasudev Lal
Vasudev Lal
Oracle
AIDeep LearningCVNLP
Shao-Yen Tseng
Shao-Yen Tseng
Intel Labs
Machine LearningNatural Language ProcessingSpeech and Audio Processing