🤖 AI Summary
Open-vocabulary video visual relation detection suffers from error propagation from object detection to relation classification in cascaded pipelines. This paper proposes the first query-driven unified framework that jointly models object detection and relation classification, enabling their co-optimization via a bidirectional mutual enhancement mechanism. Key contributions include: (1) a CLIP-based contextual refinement encoding module that enhances generalization to unseen object and relation categories; and (2) an iterative mutual enhancement module that explicitly captures semantic dependencies between objects and relations. The method integrates vision-language alignment, query-based Transformer architecture, context-aware feature refinement, and iterative representation enhancement. Evaluated on VidVRD and VidOR, it achieves state-of-the-art performance, significantly improving joint recognition accuracy for unknown-category objects and relations in open-vocabulary settings.
📝 Abstract
Open-vocabulary video visual relationship detection aims to detect objects and their relationships in videos without being restricted by predefined object or relationship categories. Existing methods leverage the rich semantic knowledge of pre-trained vision-language models such as CLIP to identify novel categories. They typically adopt a cascaded pipeline to first detect objects and then classify relationships based on the detected objects, which may lead to error propagation and thus suboptimal performance. In this paper, we propose Mutual EnhancemenT of Objects and Relationships (METOR), a query-based unified framework to jointly model and mutually enhance object detection and relationship classification in open-vocabulary scenarios. Under this framework, we first design a CLIP-based contextual refinement encoding module that extracts visual contexts of objects and relationships to refine the encoding of text features and object queries, thus improving the generalization of encoding to novel categories. Then we propose an iterative enhancement module to alternatively enhance the representations of objects and relationships by fully exploiting their interdependence to improve recognition performance. Extensive experiments on two public datasets, VidVRD and VidOR, demonstrate that our framework achieves state-of-the-art performance.