đ€ AI Summary
This work addresses the problem that interpretable concept extraction in neural networks heavily relies on sparse autoencoders (SAEs) and lacks theoretical grounding. We propose an unsupervised differential clustering method that formalizes Deleuzeâs philosophical principleââconcepts are differencesââinto a computational framework: neuron activation differences are modeled via discriminant analysis, and activation skewness weighting is introduced to enhance concept diversity. The method is cross-modal, applicable uniformly to vision, language, and audio tasks. Evaluated on five mainstream models, it yields concepts of significantly higher quality than existing unsupervised SAE variantsâapproaching supervised baselinesâwhile enabling causal intervention and effective behavioral control of neural models.
đ Abstract
We propose an alternative to sparse autoencoders (SAEs) as a simple and effective unsupervised method for extracting interpretable concepts from neural networks. The core idea is to cluster differences in activations, which we formally justify within a discriminant analysis framework. To enhance the diversity of extracted concepts, we refine the approach by weighting the clustering using the skewness of activations. The method aligns with Deleuze's modern view of concepts as differences. We evaluate the approach across five models and three modalities (vision, language, and audio), measuring concept quality, diversity, and consistency. Our results show that the proposed method achieves concept quality surpassing prior unsupervised SAE variants while approaching supervised baselines, and that the extracted concepts enable steering of a model's inner representations, demonstrating their causal influence on downstream behavior.