🤖 AI Summary
This work identifies a fundamental flaw in current LLM safety mechanisms: overreliance on explicit content filtering while neglecting user intent understanding and contextual modeling, leading to systemic bypassability. Through multi-model comparative experiments across ChatGPT, Claude, Gemini, and DeepSeek—and leveraging a prompt-engineering attack framework coupled with a human-annotated intent evaluation protocol—we empirically demonstrate that emotion-guided prompting, incremental disclosure, and academic framing reliably evade mainstream safeguards. We formally propose that *intent recognition must be an intrinsic, native capability of LLMs*, not a post-hoc filtering module; further, we find that enhanced reasoning exacerbates intent-blindness. Consequently, we advocate a paradigm shift toward *joint context-intent modeling* as the foundational principle for LLM security design—providing both theoretical grounding and actionable pathways toward robust, intent-aware safety architectures.
📝 Abstract
Current Large Language Models (LLMs) safety approaches focus on explicitly harmful content while overlooking a critical vulnerability: the inability to understand context and recognize user intent. This creates exploitable vulnerabilities that malicious users can systematically leverage to circumvent safety mechanisms. We empirically evaluate multiple state-of-the-art LLMs, including ChatGPT, Claude, Gemini, and DeepSeek. Our analysis demonstrates the circumvention of reliable safety mechanisms through emotional framing, progressive revelation, and academic justification techniques. Notably, reasoning-enabled configurations amplified rather than mitigated the effectiveness of exploitation, increasing factual precision while failing to interrogate the underlying intent. The exception was Claude Opus 4.1, which prioritized intent detection over information provision in some use cases. This pattern reveals that current architectural designs create systematic vulnerabilities. These limitations require paradigmatic shifts toward contextual understanding and intent recognition as core safety capabilities rather than post-hoc protective mechanisms.