International AI Safety Report 2025: First Key Update: Capabilities and Risk Implications

📅 2025-10-15
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Rapid advances in AI reasoning capabilities—particularly through chain-of-thought and inference-time scaling—have introduced novel cross-domain safety risks, yet existing risk analyses remain overly reliant on model scale as a proxy for capability and threat. Method: This work systematically assesses capability evolution and associated risk implications since the first international AI safety report, establishing the first systematic linkage between AI reasoning progress and emergent risks in biosafety, cybersecurity, and other domains. Using cross-domain benchmarks—including programming, mathematical reasoning, and scientific question-answering—we quantify the concurrent emergence of capability leaps and reliability deficits. Contribution/Results: We identify that non-parameter-scaling reasoning enhancements (e.g., inference-time search, self-refinement) introduce new challenges for monitoring and controllability. Our analysis revises risk assessments across critical domains, demonstrating that reasoning capability gains do not uniformly improve safety—and may in fact exacerbate certain high-stakes threats independent of parameter count.

Technology Category

Application Category

📝 Abstract
Since the publication of the first International AI Safety Report, AI capabilities have continued to improve across key domains. New training techniques that teach AI systems to reason step-by-step and inference-time enhancements have primarily driven these advances, rather than simply training larger models. As a result, general-purpose AI systems can solve more complex problems in a range of domains, from scientific research to software development. Their performance on benchmarks that measure performance in coding, mathematics, and answering expert-level science questions has continued to improve, though reliability challenges persist, with systems excelling on some tasks while failing completely on others. These capability improvements also have implications for multiple risks, including risks from biological weapons and cyber attacks. Finally, they pose new challenges for monitoring and controllability. This update examines how AI capabilities have improved since the first Report, then focuses on key risk areas where substantial new evidence warrants updated assessments.
Problem

Research questions and friction points this paper is trying to address.

Assessing AI capability improvements in reasoning and inference techniques
Evaluating reliability challenges in coding, math, and science benchmarks
Analyzing risk implications for biological weapons and cyber attacks
Innovation

Methods, ideas, or system contributions that make the work stand out.

New training techniques for step-by-step reasoning
Inference-time enhancements instead of larger models
Improved performance on coding and science benchmarks
Yoshua Bengio
Yoshua Bengio
Professor of computer science, University of Montreal, Mila, IVADO, CIFAR
Machine learningdeep learningartificial intelligence
S
Stephen Clare
Carina Prunkl
Carina Prunkl
Ethics Institute, Utrecht University
Ethics of AIGovernance of AIPhilosophy of Science and TechnologyPhilosophy of Physics
Shalaleh Rismani
Shalaleh Rismani
Postdoctoral researcher
AI Ethicsrobot ethicshuman computer interactionhuman robot interaction
Maksym Andriushchenko
Maksym Andriushchenko
ELLIS Institute Tübingen & Max Planck Institute for Intelligent Systems
AI SafetyAI AlignmentLLMsLLM agents
Ben Bucknall
Ben Bucknall
DPhil Student, University of Oxford
P
Philip Fox
Tiancheng Hu
Tiancheng Hu
University of Cambridge
natural language processingcomputational social science
C
Cameron Jones
Sam Manning
Sam Manning
Research Fellow, Centre for the Governance of AI
development economicseconomic impacts of artificial intelligence
Nestor Maslej
Nestor Maslej
Stanford University, The Stanford Institute for Human-Centered Artificial Intelligence
Artificial Intelligence
Vasilios Mavroudis
Vasilios Mavroudis
Research Scientist, Alan Turing Institute
Machine LearningSystems SecurityArtificial Intelligence
C
Conor McGlynn
M
Malcolm Murray
C
Charlotte Stix
L
Lucia Velasco
N
Nicole Wheeler
D
Daniel Privitera
Sören Mindermann
Sören Mindermann
University of Oxford, OATML
AI safetydeep learningactive learningcausal inferenceCOVID-19
Daron Acemoglu
Daron Acemoglu
Economics, MIT
T
Thomas G. Dietterich
Fredrik Heintz
Fredrik Heintz
Professor of Computer Science, Linköping University
Artificial intelligenceTrustworthy AIautonomous systemsmulti agent systemscomputational thinking
Geoffrey Hinton
Geoffrey Hinton
Emeritus Prof. Computer Science, University of Toronto
machine learningpsychologyartificial intelligencecognitive sciencecomputer science
Nick Jennings
Nick Jennings
Vice-Chancellor and President, Loughborough University
AIArtificial IntelligenceMulti-Agent SystemsIntelligent Agentsmultiagent systems
Susan Leavy
Susan Leavy
University College Dublin, Insight Centre for Data Analytics
AI EthicsArtificial IntelligenceNatural Lanugage ProcessingAlgorithmic BiasDigital Humanities