🤖 AI Summary
Existing LLM interpretability methods rely on autoencoders with implicit regularization to disentangle numerical features, but lack explicit inter-sample global sparsity guarantees—resulting in numerous low-activity dense features that undermine feature atomicity and interpretability. To address this, we propose the Binary Autoencoder (BAE), which enforces explicit, cross-sample global sparsity via minibatch-level minimum-entropy constraints, 1-bit feature discretization, and gradient estimation for backpropagation. BAE enables reliable, dynamic computation of functional entropy—a key metric for feature utility. Empirically, BAE significantly improves feature atomicity and sparsity, extracts the largest number of interpretable features in analyses of inference dynamics and in-context learning mechanisms, and substantially suppresses non-active dense features. By providing a robust, quantifiable foundation for interpreting LLM internals, BAE advances the rigor and fidelity of mechanistic interpretability research.
📝 Abstract
Existing works are dedicated to untangling atomized numerical components (features) from the hidden states of Large Language Models (LLMs) for interpreting their mechanism. However, they typically rely on autoencoders constrained by some implicit training-time regularization on single training instances (i.e., $L_1$ normalization, top-k function, etc.), without an explicit guarantee of global sparsity among instances, causing a large amount of dense (simultaneously inactive) features, harming the feature sparsity and atomization. In this paper, we propose a novel autoencoder variant that enforces minimal entropy on minibatches of hidden activations, thereby promoting feature independence and sparsity across instances. For efficient entropy calculation, we discretize the hidden activations to 1-bit via a step function and apply gradient estimation to enable backpropagation, so that we term it as Binary Autoencoder (BAE) and empirically demonstrate two major applications: (1) Feature set entropy calculation. Entropy can be reliably estimated on binary hidden activations, which we empirically evaluate and leverage to characterize the inference dynamics of LLMs and In-context Learning. (2) Feature untangling. Similar to typical methods, BAE can extract atomized features from LLM's hidden states. To robustly evaluate such feature extraction capability, we refine traditional feature-interpretation methods to avoid unreliable handling of numerical tokens, and show that BAE avoids dense features while producing the largest number of interpretable ones among baselines, which confirms the effectiveness of BAE serving as a feature extractor.