Divergences between Language Models and Human Brains

📅 2023-11-15
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) exhibit systematic deficits in social-emotional reasoning and physical commonsense representation—critical dimensions of human language understanding—despite strong linguistic performance. Method: Leveraging high-temporal-resolution magnetoencephalography (MEG) data, we establish the first LLM-driven neuro-model representational alignment framework to cross-modally localize core divergences in human–machine language processing; complemented by behavioral experiments, we formulate the “knowledge gap hypothesis” and design domain-targeted fine-tuning strategies for social/emotional and physical commonsense. Contribution/Results: Pre-fine-tuned LLMs yield significantly weaker predictions of MEG neural responses than human baselines. Domain-enhanced fine-tuning substantially improves neural predictivity, confirming that impoverished commonsense knowledge—not merely architectural or training limitations—is a fundamental bottleneck to brain-like language comprehension.
📝 Abstract
Do machines and humans process language in similar ways? Recent research has hinted at the affirmative, showing that human neural activity can be effectively predicted using the internal representations of language models (LMs). Although such results are thought to reflect shared computational principles between LMs and human brains, there are also clear differences in how LMs and humans represent and use language. In this work, we systematically explore the divergences between human and machine language processing by examining the differences between LM representations and human brain responses to language as measured by Magnetoencephalography (MEG) across two datasets in which subjects read and listened to narrative stories. Using an LLM-based data-driven approach, we identify two domains that LMs do not capture well: social/emotional intelligence and physical commonsense. We validate these findings with human behavioral experiments and hypothesize that the gap is due to insufficient representations of social/emotional and physical knowledge in LMs. Our results show that fine-tuning LMs on these domains can improve their alignment with human brain responses.
Problem

Research questions and friction points this paper is trying to address.

Language Models
Social Emotional Understanding
Physical Commonsense
Innovation

Methods, ideas, or system contributions that make the work stand out.

Neural Language Models
Social-Emotional Context
Physical Commonsense Training
🔎 Similar Papers
No similar papers found.