Shape Happens: Automatic Feature Manifold Discovery in LLMs via Supervised Multi-Dimensional Scaling

📅 2025-10-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Prior work designs geometric structures for specific features, lacking generalizability across model families and scales. Method: We propose Supervised Multidimensional Scaling (SMDS), the first method to automatically discover feature manifolds across diverse large language models (LLMs) and sizes, systematically characterizing the geometric representations of concepts in activation space. By integrating concept-labeled data into manifold learning, SMDS uncovers distinct geometric structures—circular, linear, and cluster-like—in temporal reasoning, corresponding respectively to periodicity, sequentiality, and categorical semantics. Results: We empirically demonstrate that these structures dynamically reconfigure within context and functionally support concretized reasoning. Experiments confirm their cross-model stability, interpretability, and strong correlation with downstream reasoning performance.

Technology Category

Application Category

📝 Abstract
The linear representation hypothesis states that language models (LMs) encode concepts as directions in their latent space, forming organized, multidimensional manifolds. Prior efforts focus on discovering specific geometries for specific features, and thus lack generalization. We introduce Supervised Multi-Dimensional Scaling (SMDS), a model-agnostic method to automatically discover feature manifolds. We apply SMDS to temporal reasoning as a case study, finding that different features form various geometric structures such as circles, lines, and clusters. SMDS reveals many insights on these structures: they consistently reflect the properties of the concepts they represent; are stable across model families and sizes; actively support reasoning in models; and dynamically reshape in response to context changes. Together, our findings shed light on the functional role of feature manifolds, supporting a model of entity-based reasoning in which LMs encode and transform structured representations.
Problem

Research questions and friction points this paper is trying to address.

Automatically discovering feature manifolds in language models
Revealing geometric structures of concepts in latent spaces
Analyzing how manifolds support reasoning and context adaptation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Supervised Multi-Dimensional Scaling automatically discovers feature manifolds
Method reveals geometric structures like circles and lines
Technique shows stable manifolds across model families
🔎 Similar Papers
2024-08-08International Workshop on Semantic and Social Media Adaptation and PersonalizationCitations: 13
F
Federico Tiblias
Ubiquitous Knowledge Processing Lab (UKP Lab), Department of Computer Science and Hessian Center for AI (hessian.AI), Technical University of Darmstadt
I
Irina Bigoulaeva
Ubiquitous Knowledge Processing Lab (UKP Lab), Department of Computer Science and Hessian Center for AI (hessian.AI), Technical University of Darmstadt
Jingcheng Niu
Jingcheng Niu
TU Darmstadt
Natural Language Processing
Simone Balloccu
Simone Balloccu
TU Darmstadt
NLPAIMachine LearningNLG
Iryna Gurevych
Iryna Gurevych
Full Professor, TU Darmstadt; Adjunct Professor, MBZUAI, UAE; Affiliated Professor, INSAIT, Bulgaria
Natural Language ProcessingLarge Language ModelsArtificial Intelligence