🤖 AI Summary
Large language models (LLMs) often fail to proactively avoid harmful outputs in high-stakes domains, posing critical safety risks. Method: This paper introduces InvThink—the first safety alignment framework integrating inverse reasoning. It implements a three-step inverse thinking paradigm: (1) anticipating potential failure modes, (2) reasoning backward to infer harmful consequences, and (3) generating inherently safe, avoidance-oriented responses. InvThink is optimized via joint supervised fine-tuning and reinforcement learning, enabling scalable deployment across diverse LLM families. Contribution/Results: Experiments demonstrate that InvThink reduces harmful responses by up to 15.7% over baselines (e.g., SafetyPrompt) in medical, financial, and agentic tasks, significantly mitigating the “safety tax” without degrading general reasoning capabilities. This work establishes the first systematic instantiation of safety alignment grounded in inverse reasoning, offering a novel paradigm for trustworthy AI deployment in high-risk applications.
📝 Abstract
We present InvThink, a simple yet powerful approach that gives large language models (LLMs) the capability of inverse thinking: reasoning through failure modes before generating responses. Unlike existing safety alignment methods that optimize directly for safe response, InvThink instructs models to 1) enumerate potential harms, 2) analyze their consequences, and 3) generate safe outputs that proactively avoid these risks. Our method reveals three key findings: (i) safety improvements show stronger scaling with model size compared to existing safety methods. (ii) InvThink mitigates safety tax; by training models to systematically consider failure modes, it preserves general reasoning capabilities on standard benchmarks. (iii) beyond general safety tasks, InvThink excels in high-stakes domains including external-facing (medicine, finance, law) and agentic (blackmail, murder) risk scenarios, achieving up to 15.7% reduction in harmful responses compared to baseline methods like SafetyPrompt. We further implement InvThink via supervised fine-tuning, and reinforcement learning across three LLM families. These results suggest that inverse reasoning provides a scalable and generalizable path toward safer, more capable language models.