Large Reasoning Models Are Autonomous Jailbreak Agents

📅 2025-08-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work exposes a systemic security threat: large reasoning models (LRMs) can be weaponized as autonomous jailbreaking agents against other AI systems. Addressing the challenge that conventional jailbreaking attacks are inaccessible to non-experts, we propose an autonomous jailbreaking framework grounded in multi-turn interaction, automatic planning, and systematic prompting—leveraging state-of-the-art LRMs including DeepSeek-R1, Gemini 2.5 Flash, Grok-3 Mini, and Qwen3-235B for coordinated attacks. Evaluated across 70 test cases spanning seven sensitive domains, our framework achieves a cross-model average success rate of 97.14%. Our key contribution is the first empirical demonstration that LRMs not only require robust defenses against input-based jailbreaking but also inherently function as highly effective, low-barrier jailbreaking tools. This finding necessitates a paradigm shift in AI safety—from solely defending against adversarial inputs to preventing models from being hijacked as jailbreaking enablers.

Technology Category

Application Category

📝 Abstract
Jailbreaking -- bypassing built-in safety mechanisms in AI models -- has traditionally required complex technical procedures or specialized human expertise. In this study, we show that the persuasive capabilities of large reasoning models (LRMs) simplify and scale jailbreaking, converting it into an inexpensive activity accessible to non-experts. We evaluated the capabilities of four LRMs (DeepSeek-R1, Gemini 2.5 Flash, Grok 3 Mini, Qwen3 235B) to act as autonomous adversaries conducting multi-turn conversations with nine widely used target models. LRMs received instructions via a system prompt, before proceeding to planning and executing jailbreaks with no further supervision. We performed extensive experiments with a benchmark of harmful prompts composed of 70 items covering seven sensitive domains. This setup yielded an overall attack success rate across all model combinations of 97.14%. Our study reveals an alignment regression, in which LRMs can systematically erode the safety guardrails of other models, highlighting the urgent need to further align frontier models not only to resist jailbreak attempts, but also to prevent them from being co-opted into acting as jailbreak agents.
Problem

Research questions and friction points this paper is trying to address.

LRMs simplify jailbreaking for non-experts
LRMs autonomously bypass AI safety mechanisms
LRMs erode safety guardrails of other models
Innovation

Methods, ideas, or system contributions that make the work stand out.

LRMs simplify jailbreaking via persuasive capabilities
Autonomous multi-turn conversations bypass safety mechanisms
System prompts enable unsupervised jailbreak planning
🔎 Similar Papers
No similar papers found.