π€ AI Summary
Large language models (LLMs) often fail to detect subtle safety violations in real-world applications. Method: This paper proposes the Criticality-Augmented Gatekeeper (CAG), a lightweight safety guard that distills structured, deliberative βslow thinkingβ β i.e., critical reasoning β from high-capacity LLMs via joint fine-tuning on structured critiques and safety labels. CAG employs a multitask supervised learning framework coupled with critique-guided classification. Contribution/Results: CAG introduces the first interpretable and distillable critical reasoning paradigm, achieving both high inference efficiency and significantly enhanced risk detection rigor and depth. Experiments demonstrate state-of-the-art performance across multiple safety benchmarks: CAG achieves superior F1 and AUPRC scores, outperforming LLaMA Guard 3 by +16.1% in accuracy and +27.0% in macro-F1.
π Abstract
Ensuring the safety of large language models (LLMs) is critical as they are deployed in real-world applications. Existing guardrails rely on rule-based filtering or single-pass classification, limiting their ability to handle nuanced safety violations. To address this, we propose ThinkGuard, a critique-augmented guardrail model that distills knowledge from high-capacity LLMs by generating structured critiques alongside safety labels. Fine-tuned on critique-augmented data, the captured deliberative thinking ability drastically enhances the guardrail's cautiousness and interpretability. Evaluated on multiple safety benchmarks, ThinkGuard achieves the highest average F1 and AUPRC, outperforming all baselines. Compared to LLaMA Guard 3, ThinkGuard improves accuracy by 16.1% and macro F1 by 27.0%. Moreover, it surpasses label-only fine-tuned models, confirming that structured critiques enhance both classification precision and nuanced safety reasoning while maintaining computational efficiency.