City-Scale Multi-Camera Vehicle Tracking System with Improved Self-Supervised Camera Link Model

📅 2024-05-18
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In city-scale multi-camera vehicle tracking, cross-camera trajectory matching typically relies on labor-intensive spatiotemporal annotations, leading to high deployment costs and poor scalability. Method: This paper proposes a self-supervised camera linking approach that eliminates the need for manual labels. It introduces a joint pre-matching mechanism leveraging only vehicle appearance features, trajectory-pair counts, and temporal variance to dynamically infer high-confidence spatial associations between cameras. Furthermore, it designs a probabilistic camera-linking inference framework for end-to-end camera relationship modeling. Contribution/Results: Evaluated on the CityFlow V2 benchmark, our method achieves an IDF1 score of 61.07%, setting the new state-of-the-art among fully automatic camera-linking methods. It significantly enhances system scalability and practical deployment efficiency while removing dependency on human annotation.

Technology Category

Application Category

📝 Abstract
Multi-Target Multi-Camera Tracking (MTMCT) has broad applications and forms the basis for numerous future city-wide systems (e.g. traffic management, crash detection, etc.). However, the challenge of matching vehicle trajectories across different cameras based solely on feature extraction poses significant difficulties. This article introduces an innovative multi-camera vehicle tracking system that utilizes a self-supervised camera link model. In contrast to related works that rely on manual spatial-temporal annotations, our model automatically extracts crucial multi-camera relationships for vehicle matching. The camera link is established through a pre-matching process that evaluates feature similarities, pair numbers, and time variance for high-quality tracks. This process calculates the probability of spatial linkage for all camera combinations, selecting the highest scoring pairs to create camera links. Our approach significantly improves deployment times by eliminating the need for human annotation, offering substantial improvements in efficiency and cost-effectiveness when it comes to real-world application. This pairing process supports cross camera matching by setting spatial-temporal constraints, reducing the searching space for potential vehicle matches. According to our experimental results, the proposed method achieves a new state-of-the-art among automatic camera-link based methods in CityFlow V2 benchmarks with 61.07% IDF1 Score.
Problem

Research questions and friction points this paper is trying to address.

Multi-camera vehicle tracking
Self-supervised camera link
Automatic spatial-temporal matching
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-supervised camera link model
Automatic multi-camera relationship extraction
Pre-matching process for high-quality tracks
🔎 Similar Papers
No similar papers found.
Y
Yuqiang Lin
Department of Mechanical Engineering, University of Bath, Bath, United Kingdom
S
Sam Lockyer
Department of Mechanical Engineering, University of Bath, Bath, United Kingdom
N
Nic Zhang
Department of Mechanical Engineering, University of Bath, Bath, United Kingdom
A
Adrian Evans
M
Markus Zarbock