🤖 AI Summary
Existing visual anomaly detection methods are predominantly constrained to industrial defects or synthetic scenarios, failing to generalize to the diversity and unpredictability of real-world anomalies. To address this, we propose CAVE—the first cognition-driven benchmark for real-world visual anomaly detection—grounded in cognitive science principles. CAVE introduces a fine-grained, multi-level annotation schema encompassing anomaly type, severity, and prevalence, enabling open-domain reasoning tasks such as anomaly description, explanation, and justification. Its annotation framework jointly models visual localization and human perceptual mechanisms. We systematically evaluate state-of-the-art vision-language models (VLMs) under diverse prompting strategies. Experiments reveal substantial deficiencies in current VLMs’ commonsense anomaly perception and reasoning capabilities, underscoring CAVE’s rigor and necessity as a high-challenge benchmark for advancing robust, human-aligned visual anomaly understanding.
📝 Abstract
Humans can naturally identify, reason about, and explain anomalies in their environment. In computer vision, this long-standing challenge remains limited to industrial defects or unrealistic, synthetically generated anomalies, failing to capture the richness and unpredictability of real-world anomalies. In this work, we introduce CAVE, the first benchmark of real-world visual anomalies. CAVE supports three open-ended tasks: anomaly description, explanation, and justification; with fine-grained annotations for visual grounding and categorizing anomalies based on their visual manifestations, their complexity, severity, and commonness. These annotations draw inspiration from cognitive science research on how humans identify and resolve anomalies, providing a comprehensive framework for evaluating Vision-Language Models (VLMs) in detecting and understanding anomalies. We show that state-of-the-art VLMs struggle with visual anomaly perception and commonsense reasoning, even with advanced prompting strategies. By offering a realistic and cognitively grounded benchmark, CAVE serves as a valuable resource for advancing research in anomaly detection and commonsense reasoning in VLMs.