🤖 AI Summary
This paper examines the risks and benefits of international cooperation among geopolitical rivals in technical AI safety. To address this challenge, it first systematizes AI safety subfields and constructs—novelty notwithstanding—a two-dimensional risk–value taxonomy. Drawing on historical strategic technology collaborations, empirical analysis of U.S.–China AI cooperation cases, risk categorization modeling, and comparative policy framework analysis, the study identifies low-risk, high-value collaboration avenues, including verification mechanisms and standardization protocols. It further proposes a dynamic risk governance framework tailored to AI’s distinctive technical characteristics—such as rapid iteration, dual-use potential, and opacity—and specifies prioritized cooperation pathways. The findings deliver actionable, evidence-based risk mitigation strategies for national research institutions and policymakers, thereby advancing the practical implementation of safety-oriented multilateral AI collaboration. (149 words)
📝 Abstract
International cooperation is common in AI research, including between geopolitical rivals. While many experts advocate for greater international cooperation on AI safety to address shared global risks, some view cooperation on AI with suspicion, arguing that it can pose unacceptable risks to national security. However, the extent to which cooperation on AI safety poses such risks, as well as provides benefits, depends on the specific area of cooperation. In this paper, we consider technical factors that impact the risks of international cooperation on AI safety research, focusing on the degree to which such cooperation can advance dangerous capabilities, result in the sharing of sensitive information, or provide opportunities for harm. We begin by why nations historically cooperate on strategic technologies and analyse current US-China cooperation in AI as a case study. We further argue that existing frameworks for managing associated risks can be supplemented with consideration of key risks specific to cooperation on technical AI safety research. Through our analysis, we find that research into AI verification mechanisms and shared protocols may be suitable areas for such cooperation. Through this analysis we aim to help researchers and governments identify and mitigate the risks of international cooperation on AI safety research, so that the benefits of cooperation can be fully realised.