π€ AI Summary
Existing video encoders face bottlenecks in dense prediction tasks (e.g., object tracking, semantic segmentation): image encoders (e.g., DINO, CLIP) lack temporal modeling capability, while video models (e.g., VideoMAE) yield coarse spatial representations unsuitable for pixel-level alignment. To address this, we propose FRAMEβa self-supervised frame encoder tailored for dense video understanding. FRAME takes historical and current RGB frames as input and jointly predicts DINO patch features for the current and future frames. It employs CLIP-based semantic space alignment and a lightweight temporal memory mechanism to produce token-level representations that are spatiotemporally consistent, spatially precise, and language-aligned. FRAME outperforms DINO, CLIP, and VideoMAE across seven datasets and six dense prediction tasks. Moreover, it generalizes effectively to cross-modal tasks such as video classification, achieving both compactness and strong transferability.
π Abstract
Dense video prediction tasks, such as object tracking and semantic segmentation, require video encoders that generate temporally consistent, spatially dense features for every frame. However, existing approaches fall short: image encoders like DINO or CLIP lack temporal awareness, while video models such as VideoMAE underperform compared to image encoders on dense prediction tasks. We address this gap with FRAME, a self-supervised video frame encoder tailored for dense video understanding. FRAME learns to predict current and future DINO patch features from past and present RGB frames, leading to spatially precise and temporally coherent representations. To our knowledge, FRAME is the first video encoder to leverage image-based models for dense prediction while outperforming them on tasks requiring fine-grained visual correspondence. As an auxiliary capability, FRAME aligns its class token with CLIP's semantic space, supporting language-driven tasks such as video classification. We evaluate FRAME across six dense prediction tasks on seven datasets, where it consistently outperforms image encoders and existing self-supervised video models. Despite its versatility, FRAME maintains a compact architecture suitable for a range of downstream applications.