π€ AI Summary
Existing unsupervised video object-centric learning (OCL) methods suffer from two key limitations: (i) neglecting next-frame featuresβthe most discriminative source for query prediction; and (ii) failing to explicitly model the dynamic knowledge underlying object state transitions. To address these, we propose RandSF.Q, the first method to introduce random slot-feature pairs into the attention-based query mechanism. By randomly sampling training transitions, RandSF.Q explicitly learns inter-frame state dynamics while incorporating next-frame features to enhance query representation. Our approach employs an attention-driven slot update architecture that jointly leverages random sampling and feature aggregation, enabling precise object segmentation and dynamic modeling under fully unsupervised conditions. On standard object discovery benchmarks, RandSF.Q significantly outperforms prior state-of-the-art methods, achieving up to a 10-percentage-point improvement and establishing new SOTA performance.
π Abstract
Unsupervised video Object-Centric Learning (OCL) is promising as it enables object-level scene representation and dynamics modeling as we humans do. Mainstream video OCL methods adopt a recurrent architecture: An aggregator aggregates current video frame into object features, termed slots, under some queries; A transitioner transits current slots to queries for the next frame. This is an effective architecture but all existing implementations both ( extit{i1}) neglect to incorporate next frame features, the most informative source for query prediction, and ( extit{i2}) fail to learn transition dynamics, the knowledge essential for query prediction. To address these issues, we propose Random Slot-Feature pair for learning Query prediction (RandSF.Q): ( extit{t1}) We design a new transitioner to incorporate both slots and features, which provides more information for query prediction; ( extit{t2}) We train the transitioner to predict queries from slot-feature pairs randomly sampled from available recurrences, which drives it to learn transition dynamics. Experiments on scene representation demonstrate that our method surpass existing video OCL methods significantly, e.g., up to 10 points on object discovery, setting new state-of-the-art. Such superiority also benefits downstream tasks like dynamics modeling. Our core source code and training logs are available as the supplement.