Are LLMs Smarter Than Chimpanzees? An Evaluation on Perspective Taking and Knowledge State Estimation

📅 2026-01-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates whether large language models (LLMs) possess the capacity to infer others’ knowledge states and intentions—a core component of theory of mind that distinguishes humans from chimpanzees. To this end, it introduces, for the first time, the knowledge-state tracking paradigm from cognitive anthropology into LLM evaluation, designing two story-comprehension tasks: detecting inconsistencies between a character’s actions and their subjective knowledge, and predicting subsequent behavior based on such inferences. Experimental results demonstrate that mainstream LLMs perform near chance levels on both tasks, significantly underperforming human participants. These findings reveal a profound deficit in LLMs’ foundational theory-of-mind capabilities and establish a novel evaluation paradigm and benchmark for assessing cognitive competencies in artificial systems.

Technology Category

Application Category

📝 Abstract
Cognitive anthropology suggests that the distinction of human intelligence lies in the ability to infer other individuals'knowledge states and understand their intentions. In comparison, our closest animal relative, chimpanzees, lack the capacity to do so. With this paper, we aim to evaluate LLM performance in the area of knowledge state tracking and estimation. We design two tasks to test (1) if LLMs can detect when story characters, through their actions, demonstrate knowledge they should not possess, and (2) if LLMs can predict story characters'next actions based on their own knowledge vs. objective truths they do not know. Results reveal that most current state-of-the-art LLMs achieve near-random performance on both tasks, and are substantially inferior to humans. We argue future LLM research should place more weight on the abilities of knowledge estimation and intention understanding.
Problem

Research questions and friction points this paper is trying to address.

perspective taking
knowledge state estimation
large language models
theory of mind
cognitive anthropology
Innovation

Methods, ideas, or system contributions that make the work stand out.

knowledge state estimation
perspective taking
theory of mind
large language models
cognitive evaluation
🔎 Similar Papers
No similar papers found.
Dingyi Yang
Dingyi Yang
Renmin University of China
Natural Language ProcessingVision and LanguageGenerationEvaluation
J
Junqi Zhao
College of Computing and Data Science, Nanyang Technological University
X
Xue Li
University of Science and Technology of China
Ce Li
Ce Li
CUMTB
Video UnderstandingBehavior AnalysisEvent Detection
B
Boyang Li
College of Computing and Data Science, Nanyang Technological University