🤖 AI Summary
To address background shift and catastrophic forgetting in incremental object detection—caused by background class overlap across tasks—this paper proposes the Class-Agnostic Shared Attribute (CASA) framework. CASA is the first to incorporate generic semantic attributes from vision-language foundation models into incremental detection, enabling class-agnostic, decoupled attribute representations. It introduces a dynamic attribute selection-and-freezing mechanism alongside an attribute weighting matrix to support adaptive evolution of the semantic space. Built upon OWL-ViT, CASA integrates LLM-generated textual attributes, parameter-efficient fine-tuning (PEFT), and attribute importance modeling. Evaluated on COCO under both two-stage and multi-stage incremental settings, CASA achieves state-of-the-art performance with only a 0.7% parameter increase, significantly enhancing model scalability and cross-task generalization capability.
📝 Abstract
Incremental object detection (IOD) is challenged by background shift, where background categories in sequential data may include previously learned or future classes. Inspired by the vision-language foundation models such as CLIP, these models capture shared attributes from extensive image-text paired data during pre-training. We propose a novel method utilizing attributes in vision-language foundation models for incremental object detection. Our method constructs a Class-Agnostic Shared Attribute base (CASA) to capture common semantic information among incremental classes. Specifically, we utilize large language models to generate candidate textual attributes and select the most relevant ones based on current training data, recording their significance in an attribute assignment matrix. For subsequent tasks, we freeze the retained attributes and continue selecting from the remaining candidates while updating the attribute assignment matrix accordingly. Furthermore, we employ OWL-ViT as our baseline, preserving the original parameters of the pre-trained foundation model. Our method adds only 0.7% to parameter storage through parameter-efficient fine-tuning to significantly enhance the scalability and adaptability of IOD. Extensive two-phase and multi-phase experiments on the COCO dataset demonstrate the state-of-the-art performance of our proposed method.