In Which Areas of Technical AI Safety Could Geopolitical Rivals Cooperate?

📅 2025-04-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper examines the risks and benefits of international cooperation among geopolitical rivals in technical AI safety. To address this challenge, it first systematizes AI safety subfields and constructs—novelty notwithstanding—a two-dimensional risk–value taxonomy. Drawing on historical strategic technology collaborations, empirical analysis of U.S.–China AI cooperation cases, risk categorization modeling, and comparative policy framework analysis, the study identifies low-risk, high-value collaboration avenues, including verification mechanisms and standardization protocols. It further proposes a dynamic risk governance framework tailored to AI’s distinctive technical characteristics—such as rapid iteration, dual-use potential, and opacity—and specifies prioritized cooperation pathways. The findings deliver actionable, evidence-based risk mitigation strategies for national research institutions and policymakers, thereby advancing the practical implementation of safety-oriented multilateral AI collaboration. (149 words)

Technology Category

Application Category

📝 Abstract
International cooperation is common in AI research, including between geopolitical rivals. While many experts advocate for greater international cooperation on AI safety to address shared global risks, some view cooperation on AI with suspicion, arguing that it can pose unacceptable risks to national security. However, the extent to which cooperation on AI safety poses such risks, as well as provides benefits, depends on the specific area of cooperation. In this paper, we consider technical factors that impact the risks of international cooperation on AI safety research, focusing on the degree to which such cooperation can advance dangerous capabilities, result in the sharing of sensitive information, or provide opportunities for harm. We begin by why nations historically cooperate on strategic technologies and analyse current US-China cooperation in AI as a case study. We further argue that existing frameworks for managing associated risks can be supplemented with consideration of key risks specific to cooperation on technical AI safety research. Through our analysis, we find that research into AI verification mechanisms and shared protocols may be suitable areas for such cooperation. Through this analysis we aim to help researchers and governments identify and mitigate the risks of international cooperation on AI safety research, so that the benefits of cooperation can be fully realised.
Problem

Research questions and friction points this paper is trying to address.

Assessing risks and benefits of international AI safety cooperation
Identifying suitable areas for geopolitical rivals to collaborate on AI safety
Mitigating risks in US-China AI safety research collaboration
Innovation

Methods, ideas, or system contributions that make the work stand out.

Analyze US-China AI cooperation case study
Supplement risk frameworks for AI safety
Identify AI verification mechanisms for cooperation
Ben Bucknall
Ben Bucknall
DPhil Student, University of Oxford
S
Saad Siddiqui
Safe AI Forum & Oxford Martin AI Governance Initiative
Ben Harack
Ben Harack
Oxford University Department of Politics and International Relations
International RelationsArtificial IntelligenceAI GovernanceSemiconductors
Anka Reuel
Anka Reuel
CS Ph.D. Candidate, Stanford University
AI GovernanceResponsible AIAI EthicsAI Safety
Patricia Paskov
Patricia Paskov
RAND
AI evaluationAI governanceeconomicsinternational development
Casey Mahoney
Casey Mahoney
PhD Candidate, University of Pennsylvania
Political scienceinternational relationstechnology
S
Soren Mindermann
Mila - Quebec AI Institute
S
Scott Singer
Carnegie Endowment for International Peace & Oxford Martin AI Governance Initiative
V
Vinay Hiremath
Centre for the Governance of AI
C
Charbel-Raphaël Ségerie
Centre pour la Sécurité de l’IA (CeSIA)
O
Oscar Delaney
Institute for AI Policy and Strategy
Alessandro Abate
Alessandro Abate
Professor of Verification and Control, University of Oxford, UK
Formal VerificationControl TheoryStochastic Hybrid SystemsCyber-Physical SystemsEnergy and Safety-Critical Systems
Fazl Barez
Fazl Barez
University of Oxford
AI SafetyExplainabilityInterpretabilityAI Governance and Policy
M
Michael K. Cohen
UC Berkeley & Center for Human Compatible AI
Philip Torr
Philip Torr
Professor, University of Oxford
Department of Engineering
Ferenc Huszár
Ferenc Huszár
University of Cambridge
Machine LearningDeep LearningArtificial IntelligenceCausal InferenceBayesian Reasoning
Anisoara Calinescu
Anisoara Calinescu
Department of Computer Science & Reuben College, University of Oxford
Complex SystemsEmergenceEntropyAgent-Based ModellingSupply Chain and Manufacturing systems
Gabriel Davis Jones
Gabriel Davis Jones
University of Oxford
Maternal and Neonatal HealthNeuroscienceComputer ScienceArtifical IntelligenceGlobal Health
Yoshua Bengio
Yoshua Bengio
Professor of computer science, University of Montreal, Mila, IVADO, CIFAR
Machine learningdeep learningartificial intelligence
Robert Trager
Robert Trager
University of Oxford
AI GovernanceDiplomacyInstitutional DesignSocial TheoryApplied Mathematics