🤖 AI Summary
AI cognitive partners are evolving from mere tools into novel collaborators in human cognition, introducing real-time, individual-level, and societal risks that transcend those associated with conventional AI tools and agents. To address this, we propose RISc—the first multi-level risk analysis framework specifically designed for collaborative cognition. RISc systematically identifies emergent risk categories, constructs quantifiable assessment metrics, and devises tiered governance strategies. Integrating insights from cognitive science, human–AI interaction, AI safety, and policy analysis, the framework employs qualitative modeling, a novel risk taxonomy, and cross-scale evaluation methods. Key contributions include: (1) a structured risk identification checklist; (2) an operational, metrics-based assessment system; and (3) tiered mitigation guidelines tailored for developers and policymakers. Collectively, these advances provide both theoretical foundations and actionable pathways for building safe, controllable AI-augmented cognitive ecosystems. (149 words)
📝 Abstract
Artificial Intelligence (AI) systems have historically been used as tools that execute narrowly defined tasks. Yet recent advances in AI have unlocked possibilities for a new class of models that genuinely collaborate with humans in complex reasoning, from conceptualizing problems to brainstorming solutions. Such AI thought partners enable novel forms of collaboration and extended cognition, yet they also pose major risks-including and beyond risks of typical AI tools and agents. In this commentary, we systematically identify risks of AI thought partners through a novel framework that identifies risks at multiple levels of analysis, including Real-time, Individual, and Societal risks arising from collaborative cognition (RISc). We leverage this framework to propose concrete metrics for risk evaluation, and finally suggest specific mitigation strategies for developers and policymakers. As AI thought partners continue to proliferate, these strategies can help prevent major harms and ensure that humans actively benefit from productive thought partnerships.