Do Large Language Models Mirror Cognitive Language Processing?

πŸ“… 2024-02-28
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 2
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Understanding the neural alignment between large language models (LLMs) and human language processing remains a fundamental challenge in cognitive neuroscience and AI. Method: We systematically investigated representational alignment between 23 state-of-the-art LLMs and fMRI-measured neural responses in human language cortex using representational similarity analysis (RSA). Contribution/Results: We provide the first causal characterization of how pretraining data scale, model scaling, alignment-specific fine-tuning, and prompt design affect LLM–brain correspondence. Pretraining scale and supervised alignment training significantly enhance alignment; explicit semantic prompts improve cross-subject consistency, whereas noisy prompts degrade it; and benchmark scores (e.g., MMLU) strongly correlate with neural alignment (r > 0.8). These findings establish neural alignment as an objective, quantifiable metric for assessing the cognitive interpretability of LLMs, offering a principled evaluation framework for brain-inspired AI development.

Technology Category

Application Category

πŸ“ Abstract
Large Language Models (LLMs) have demonstrated remarkable abilities in text comprehension and logical reasoning, indicating that the text representations learned by LLMs can facilitate their language processing capabilities. In neuroscience, brain cognitive processing signals are typically utilized to study human language processing. Therefore, it is natural to ask how well the text embeddings from LLMs align with the brain cognitive processing signals, and how training strategies affect the LLM-brain alignment? In this paper, we employ Representational Similarity Analysis (RSA) to measure the alignment between 23 mainstream LLMs and fMRI signals of the brain to evaluate how effectively LLMs simulate cognitive language processing. We empirically investigate the impact of various factors (e.g., pre-training data size, model scaling, alignment training, and prompts) on such LLM-brain alignment. Experimental results indicate that pre-training data size and model scaling are positively correlated with LLM-brain similarity, and alignment training can significantly improve LLM-brain similarity. Explicit prompts contribute to the consistency of LLMs with brain cognitive language processing, while nonsensical noisy prompts may attenuate such alignment. Additionally, the performance of a wide range of LLM evaluations (e.g., MMLU, Chatbot Arena) is highly correlated with the LLM-brain similarity.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Neural Similarity
Training Methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Representational Similarity Analysis (RSA)
Large Language Models (LLMs) vs. Human Brain
Training Parameters Impact
πŸ”Ž Similar Papers
No similar papers found.
Y
Yuqi Ren
College of Intelligence and Computing, Tianjin University
Renren Jin
Renren Jin
College of Intelligence and Computing, Tianjin University
Natural Language Processing
T
Tongxuan Zhang
College of Computer and Information Engineering, Tianjin Normal University
Deyi Xiong
Deyi Xiong
Professor, College of Intelligence and Computing, Tianjin University, China
Natural Language ProcessingLarge Language ModelsAI4Science