🤖 AI Summary
To address the high annotation cost and low efficiency in training object detection models, this paper proposes two active sampling strategies grounded in the Lipschitz continuity assumption: (1) uniform sampling to ensure comprehensive coverage of the state space, and (2) frame-difference sampling to model temporal redundancy in videos and identify frames with high information gain. This work is the first to systematically incorporate Lipschitz continuity theory into detection data sampling design, establishing an interpretable and analytically tractable sampling framework; it also introduces the novel frame-difference sampling paradigm, substantially reducing reliance on dense manual annotations. Evaluated on COCO and a custom video object detection dataset, the proposed methods achieve up to 62–78% annotation reduction and ~45% fewer training iterations, while maintaining mAP degradation below 0.5%, demonstrating the feasibility of high-quality, efficient training under small-sample regimes.
📝 Abstract
Two sampling strategies are investigated to enhance efficiency in training a deep learning object detection model. These sampling strategies are employed under the assumption of Lipschitz continuity of deep learning models. The first strategy is uniform sampling which seeks to obtain samples evenly yet randomly through the state space of the object dynamics. The second strategy of frame difference sampling is developed to explore the temporal redundancy among successive frames in a video. Experiment result indicates that these proposed sampling strategies provide a dataset that yields good training performance while requiring relatively few manually labelled samples.