A Definition of AGI

📅 2025-10-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current AGI research lacks an operational definition, hindering quantitative assessment of the cognitive gap between AI systems and humans. To address this, we propose the first theoretically grounded, ten-dimensional AGI evaluation framework, derived from the Cattell–Horn–Carroll (CHC) theory of human cognition, covering core domains including reasoning, memory, and perception. Leveraging standardized psychometric paradigms, we conduct cross-domain, comprehensive benchmarking of leading large language models, yielding fine-grained, “sawtooth”-shaped cognitive profiles. Results reveal stark imbalances: GPT-4 scores 27% on the AGI metric, while GPT-5 achieves 58%, with pronounced deficits in long-term memory and other foundational mechanisms. This work establishes the first theory-driven, quantifiable AGI benchmark and empirically demonstrates the severe heterogeneity in current models’ cognitive capabilities—providing both a rigorous evaluation standard and actionable insights for guiding AGI development.

Technology Category

Application Category

📝 Abstract
The lack of a concrete definition for Artificial General Intelligence (AGI) obscures the gap between today's specialized AI and human-level cognition. This paper introduces a quantifiable framework to address this, defining AGI as matching the cognitive versatility and proficiency of a well-educated adult. To operationalize this, we ground our methodology in Cattell-Horn-Carroll theory, the most empirically validated model of human cognition. The framework dissects general intelligence into ten core cognitive domains-including reasoning, memory, and perception-and adapts established human psychometric batteries to evaluate AI systems. Application of this framework reveals a highly "jagged" cognitive profile in contemporary models. While proficient in knowledge-intensive domains, current AI systems have critical deficits in foundational cognitive machinery, particularly long-term memory storage. The resulting AGI scores (e.g., GPT-4 at 27%, GPT-5 at 58%) concretely quantify both rapid progress and the substantial gap remaining before AGI.
Problem

Research questions and friction points this paper is trying to address.

Defining AGI as human-level cognitive versatility
Quantifying AI progress through cognitive domain evaluation
Identifying critical deficits in current AI systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Framework quantifies AGI using human cognition theory
Adapts psychometric tests to evaluate AI cognitive domains
Reveals AI's jagged profile with quantified progress gaps
🔎 Similar Papers
No similar papers found.
Dan Hendrycks
Dan Hendrycks
Director of the Center for AI Safety (advisor for xAI and Scale)
AI SafetyML Reliability
Dawn Song
Dawn Song
Professor of Computer Science, UC Berkeley
Computer Security and Privacy
Christian Szegedy
Christian Szegedy
Researcher
Deep learningFormal reasoning
Honglak Lee
Honglak Lee
LG AI Research / U. Michigan
Machine LearningDeep LearningReinforcement LearningComputer VisionArtificial Intelligence
Yarin Gal
Yarin Gal
Professor of Machine Learning, University of Oxford
Machine LearningArtificial IntelligenceProbability TheoryStatistics
Erik Brynjolfsson
Erik Brynjolfsson
Professor at Stanford; NBER; Stanford Digital Economy Lab
EconomicsInformation EconomicsEconomics of AIProductivityIntangible Assets
Sharon Li
Sharon Li
University of Wisconsin-Madison
Machine learningReliable AI
Andy Zou
Andy Zou
PhD Student, Carnegie Mellon University
ML SafetyAI Safety
Lionel Levine
Lionel Levine
Professor, Cornell University
probabilitycombinatoricsstatistical mechanicsAI safety
B
Bo Han
Hong Kong Baptist University
J
Jie Fu
HKUST
Ziwei Liu
Ziwei Liu
Associate Professor, Nanyang Technological University
Computer VisionMachine LearningComputer Graphics
Jinwoo Shin
Jinwoo Shin
ICT Endowed Chair Professor
Machine LearningDeep Learning
Kimin Lee
Kimin Lee
KAIST
Artificial IntelligenceReinforcement LearningDeep Learning
Mantas Mazeika
Mantas Mazeika
Center for AI Safety
ML SafetyAI SafetyMachine EthicsML Reliability
Long Phan
Long Phan
Center for AI Safety
LLMAI Safety
G
George Ingebretsen
Center for AI Safety
A
Adam Khoja
Center for AI Safety
Cihang Xie
Cihang Xie
Assistant Professor, University of California, Santa Cruz
Computer VisionMachine Learning
Olawale Salaudeen
Olawale Salaudeen
Postdoctoral Associate, MIT
Trustworthy AIAI for SocietyDistribution ShiftCausality
Matthias Hein
Matthias Hein
Professor of Computer Science, University of Tübingen
Machine LearningOptimizationStatistics
K
Kevin Zhao
University of Washington
Alexander Pan
Alexander Pan
UC Berkeley
artificial intelligencemachine learning
David Duvenaud
David Duvenaud
Associate Professor, University of Toronto
LLM EvalsDifferential EquationsApproximate Inference
B
Bo Li
University of Chicago