π€ AI Summary
Existing single-object tracking methods suffer from limitations in reference-video modality combinations (typically restricted to a single reference modality), fragmented model architectures, and poor generalization. To address these issues, this paper proposes the first unified multi-modal single-object tracking framework. It supports arbitrary combinations of three reference modalities (RGB, infrared, event) and four video modalities (RGB, infrared, event, depth). The framework employs a shared Transformer backbone, cross-modal feature alignment, and dynamic prompt fusion to enable parameter-efficient sharing and joint cross-modal inference. Evaluated on 18 benchmarks, it achieves state-of-the-art performance: +3.0% AUC on TNL2K and outperforms Un-Track by over 2.0% on RGB+X tasksβthe primary metric for cross-modal tracking. These results demonstrate significantly enhanced model generality and deployment flexibility in real-world multi-sensor scenarios.
π Abstract
Single object tracking aims to localize target object with specific reference modalities (bounding box, natural language or both) in a sequence of specific video modalities (RGB, RGB+Depth, RGB+Thermal or RGB+Event.). Different reference modalities enable various human-machine interactions, and different video modalities are demanded in complex scenarios to enhance tracking robustness. Existing trackers are designed for single or several video modalities with single or several reference modalities, which leads to separate model designs and limits practical applications. Practically, a unified tracker is needed to handle various requirements. To the best of our knowledge, there is still no tracker that can perform tracking with these above reference modalities across these video modalities simultaneously. Thus, in this paper, we present a unified tracker, UniSOT, for different combinations of three reference modalities and four video modalities with uniform parameters. Extensive experimental results on 18 visual tracking, vision-language tracking and RGB+X tracking benchmarks demonstrate that UniSOT shows superior performance against modality-specific counterparts. Notably, UniSOT outperforms previous counterparts by over 3.0% AUC on TNL2K across all three reference modalities and outperforms Un-Track by over 2.0% main metric across all three RGB+X video modalities.