Knowing Your Nonlinearities: Shapley Interactions Reveal the Underlying Structure of Data

📅 2024-03-19
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates how multimodal models—spanning language, speech, and vision—encode structured input interactions (e.g., syntactic dependencies, coarticulation, object boundaries) within their internal representations. To address this, we propose an interpretability framework grounded in the Shapley–Taylor Interaction Index (STII), enabling the first cross-modal validation of interaction patterns. Applying STII across diverse architectures, we find: (i) in masked and autoregressive language models, nonlinear interactions are significantly modulated by syntactic distance; (ii) in speech recognition models, interaction structures align with articulatory-physiological constraints; and (iii) in CNN- and Transformer-based image classifiers, higher-order interactions naturally concentrate at semantic boundaries. Collectively, these results demonstrate that STII robustly captures modality-specific structural priors, offering a unified, empirically verifiable metric for probing representational geometry across modalities.

Technology Category

Application Category

📝 Abstract
Measuring nonlinear feature interaction is an established approach to understanding complex patterns of attribution in many models. In this paper, we use Shapley Taylor interaction indices (STII) to analyze the impact of underlying data structure on model representations in a variety of modalities, tasks, and architectures. Considering linguistic structure in masked and auto-regressive language models (MLMs and ALMs), we find that STII increases within idiomatic expressions and that MLMs scale STII with syntactic distance, relying more on syntax in their nonlinear structure than ALMs do. Our speech model findings reflect the phonetic principal that the openness of the oral cavity determines how much a phoneme varies based on its context. Finally, we study image classifiers and illustrate that feature interactions intuitively reflect object boundaries. Our wide range of results illustrates the benefits of interdisciplinary work and domain expertise in interpretability research.
Problem

Research questions and friction points this paper is trying to address.

Understand how language models represent linguistic structure internally
Measure pairwise Shapley interactions to decode model input relationships
Analyze interaction patterns linked to syntax, semantics, and phonetics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Using Shapley Taylor indices for model analysis
Measuring pairwise interactions in language models
Relating interactions to linguistic structures
D
Divyansh Singhvi
IIT Kanpur
Andrej Erkelens
Andrej Erkelens
Independent researcher
Raghav Jain
Raghav Jain
PhD in CS at UCSD
NLPLLMsSocial Networks
D
Diganta Misra
Carnegie Mellon University
N
Naomi Saphra
Harvard University