๐ค AI Summary
To address the high computational cost and poor real-time performance of dense, long-term 3D point tracking in videos, this paper proposes an efficient tracking framework. Methodologically, it introduces a coarse-to-fine iterative tracking strategy, an end-to-end learnable Transformer-based interpolation module, and optimized feature extraction and matching procedures; additionally, a progressive trajectory expansion mechanism is incorporated to significantly reduce redundant computations. The core contribution lies in formulating interpolation as a learnable spatiotemporal relational reasoning task, jointly optimizing feature representation and the tracking pipeline. Extensive experiments demonstrate that the framework achieves state-of-the-art (SOTA) accuracy while accelerating inference by 5โ100ร over prior methodsโenabling, for the first time, millisecond-level dense long-term 3D point tracking. This breakthrough establishes a new paradigm for real-time applications such as AR/VR and robotic vision.
๐ Abstract
We propose a novel algorithm for accelerating dense long-term 3D point tracking in videos. Through analysis of existing state-of-the-art methods, we identify two major computational bottlenecks. First, transformer-based iterative tracking becomes expensive when handling a large number of trajectories. To address this, we introduce a coarse-to-fine strategy that begins tracking with a small subset of points and progressively expands the set of tracked trajectories. The newly added trajectories are initialized using a learnable interpolation module, which is trained end-to-end alongside the tracking network. Second, we propose an optimization that significantly reduces the cost of correlation feature computation, another key bottleneck in prior methods. Together, these improvements lead to a 5-100x speedup over existing approaches while maintaining state-of-the-art tracking accuracy.