🤖 AI Summary
To address degraded action recognition performance in monocular videos caused by severe occlusion and cluttered scenes, this work introduces large language model (LLM)-encoded commonsense priors explicitly into the action recognition pipeline for the first time. Methodologically, we propose a three-stage framework: (1) video context summarization; (2) prompt-enhanced, commonsense-driven scene text description modeling; and (3) vision–language cross-modal fusion with sequence-level commonsense reasoning. This design enables explicit modeling of implicit contextual cues—including human–object interactions, object functionality, and activity logic—thereby significantly improving occlusion robustness. Experiments on Action Genome and Charades demonstrate that our method achieves average accuracy gains of 5.2%–8.7% on occlusion-heavy subsets, validating the effectiveness of integrating structured commonsense knowledge for complex action understanding.
📝 Abstract
Recent video action recognition methods have shown excellent performance by adapting large-scale pre-trained language-image models to the video domain. However, language models contain rich common sense priors - the scene contexts that humans use to constitute an understanding of objects, human-object interactions, and activities - that have not been fully exploited. In this paper, we introduce a framework incorporating language-driven common sense priors to identify cluttered video action sequences from monocular views that are often heavily occluded. We propose: (1) A video context summary component that generates candidate objects, activities, and the interactions between objects and activities; (2) A description generation module that describes the current scene given the context and infers subsequent activities, through auxiliary prompts and common sense reasoning; (3) A multi-modal activity recognition head that combines visual and textual cues to recognize video actions. We demonstrate the effectiveness of our approach on the challenging Action Genome and Charades datasets.