Exploring Similarity between Neural and LLM Trajectories in Language Processing

📅 2025-09-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates representational similarities between large language models (LLMs) and the human brain during language processing. We propose a dynamic neuro-model trajectory alignment framework that integrates ridge regression with multidimensional trajectory features—magnitude, angular dynamics, and uncertainty—to systematically compare hidden-state representations from 16 multilingual LLMs against EEG-derived neural responses. Our analysis reveals, for the first time, a significant correspondence between high-layer LLM representations and the N400 event-related potential, identifying key network layers underlying semantic integration. Crucially, we uncover a fundamental mechanistic divergence: human cortical processing relies on continuous, iterative computation, whereas LLMs exhibit discrete, stepwise activation patterns. These findings replicate robustly across English and Chinese language tasks. The work establishes a novel interdisciplinary paradigm bridging AI interpretability and cognitive neuroscience, providing empirically grounded insights into the neural plausibility of language model architectures.

Technology Category

Application Category

📝 Abstract
Understanding the similarity between large language models (LLMs) and human brain activity is crucial for advancing both AI and cognitive neuroscience. In this study, we provide a multilinguistic, large-scale assessment of this similarity by systematically comparing 16 publicly available pretrained LLMs with human brain responses during natural language processing tasks in both English and Chinese. Specifically, we use ridge regression to assess the representational similarity between LLM embeddings and electroencephalography (EEG) signals, and analyze the similarity between the "neural trajectory" and the "LLM latent trajectory." This method captures key dynamic patterns, such as magnitude, angle, uncertainty, and confidence. Our findings highlight both similarities and crucial differences in processing strategies: (1) We show that middle-to-high layers of LLMs are central to semantic integration and correspond to the N400 component observed in EEG; (2) The brain exhibits continuous and iterative processing during reading, whereas LLMs often show discrete, stage-end bursts of activity, which suggests a stark contrast in their real-time semantic processing dynamics. This study could offer new insights into LLMs and neural processing, and also establish a critical framework for future investigations into the alignment between artificial intelligence and biological intelligence.
Problem

Research questions and friction points this paper is trying to address.

Compares neural trajectories and LLM latent dynamics in language processing
Analyzes EEG-LLM similarity across 16 models and two languages
Identifies differences in continuous brain vs discrete LLM processing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Used ridge regression for neural-LLM similarity analysis
Compared neural and LLM trajectories across multiple languages
Analyzed dynamic patterns like magnitude, angle and uncertainty
🔎 Similar Papers
No similar papers found.
Xin Xiao
Xin Xiao
ByteDance Research
VLAVLM
K
Kaiwen Wei
School of Computer Science, Chongqing University, Chongqing, China
J
Jiang Zhong
School of Computer Science, Chongqing University, Chongqing, China
D
Dongshuo Yin
Department of Computer Science and Technology, Tsinghua University, Beijing, China
Y
Yu Tian
Dept. of Comp. Sci. and Tech., Institute for AI, Tsinghua University, Beijing, China
X
Xuekai Wei
School of Computer Science, Chongqing University, Chongqing, China
M
Mingliang Zhou
School of Computer Science, Chongqing University, Chongqing, China