🤖 AI Summary
This work addresses the limitations of vision-language models (VLMs) in situational awareness—specifically, their poor recognition of infrequent critical events, insufficient detail capture, and low output reliability—by proposing an enhanced framework that integrates traditional computer vision with explicit logical reasoning. The approach introduces fine-grained event parsing and a logic-guided, intelligent fine-tuning strategy, while also generating interpretable justifications for the first time during inference. This significantly improves both the accuracy of rare-event recognition and the trustworthiness of model outputs. By coupling discriminative capabilities with transparent, traceable reasoning chains, the method not only boosts VLM performance but also provides a verifiable basis for validating or challenging its conclusions.
📝 Abstract
Vision-Language Models (VLMs) offer the ability to generate high-level, interpretable descriptions of complex activities from images and videos, making them valuable for situational awareness (SA) applications. In such settings, the focus is on identifying infrequent but significant events with high reliability and accuracy, while also extracting fine-grained details and assessing recognition quality. In this paper, we propose an approach that integrates VLMs with traditional computer vision methods through explicit logic reasoning to enhance SA in three key ways: (a) extracting fine-grained event details, (b) employing an intelligent fine-tuning (FT) strategy that achieves substantially higher accuracy than uninformed selection, and (c) generating justifications for VLM outputs during inference. We demonstrate that our intelligent FT mechanism improves the accuracy and provides a valuable means, during inferencing, to either confirm the validity of the VLM output or indicate why it may be questionable.