🤖 AI Summary
Large Vision-Language Models (LVLMs) are vulnerable to adversarial jailbreaking attacks via seemingly benign, obfuscated prompts. Method: This paper proposes a lightweight, model-agnostic preprocessing defense framework that integrates fine-grained, multi-class safety classification with category-specific response strategies—namely, blocking, reconstruction, or forwarding—guided by a safety classifier and modular decision logic applied at inference time. The approach performs real-time detection and adaptive intervention on inputs without fine-tuning or modifying the target LVLM. Contribution/Results: Evaluated across five benchmarks and five state-of-the-art LVLMs, the method significantly reduces jailbreaking success rates and instruction deviation while fully preserving original task performance. It incurs zero computational overhead, offers plug-and-play deployment, and supports flexible extension to novel attack types—delivering high-compatibility, multimodal safety protection.
📝 Abstract
Large Vision-Language Models (LVLMs) unlock powerful multimodal reasoning but also expand the attack surface, particularly through adversarial inputs that conceal harmful goals in benign prompts. We propose SHIELD, a lightweight, model-agnostic preprocessing framework that couples fine-grained safety classification with category-specific guidance and explicit actions (Block, Reframe, Forward). Unlike binary moderators, SHIELD composes tailored safety prompts that enforce nuanced refusals or safe redirection without retraining. Across five benchmarks and five representative LVLMs, SHIELD consistently lowers jailbreak and non-following rates while preserving utility. Our method is plug-and-play, incurs negligible overhead, and is easily extendable to new attack types -- serving as a practical safety patch for both weakly and strongly aligned LVLMs.