BiggerGait: Unlocking Gait Recognition with Layer-wise Representations from Large Vision Models

📅 2025-05-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing large vision model (LVM)-based gait recognition methods heavily rely on handcrafted gait priors, overlooking the intrinsic potential of LVMs’ hierarchical representations. This work proposes a general baseline framework that requires no strong gait-specific priors and, for the first time, systematically reveals the cross-layer complementarity of intermediate-layer representations in LVMs for gait recognition. Our method leverages layered fusion, multi-source transfer learning, feature disentanglement, and ensemble strategies to efficiently exploit LVMs’ hierarchical semantic structure. Evaluated on four major benchmarks—CCPG, CAISA-B*, SUSTech1K, and CCGR_MINI—the approach significantly outperforms prior LVM-based methods, demonstrating superior cross-domain robustness and generalization capability. It establishes a new paradigm for fine-grained behavioral recognition driven by LVMs, shifting focus from task-specific priors to principled utilization of inherent model semantics.

Technology Category

Application Category

📝 Abstract
Large vision models (LVM) based gait recognition has achieved impressive performance. However, existing LVM-based approaches may overemphasize gait priors while neglecting the intrinsic value of LVM itself, particularly the rich, distinct representations across its multi-layers. To adequately unlock LVM's potential, this work investigates the impact of layer-wise representations on downstream recognition tasks. Our analysis reveals that LVM's intermediate layers offer complementary properties across tasks, integrating them yields an impressive improvement even without rich well-designed gait priors. Building on this insight, we propose a simple and universal baseline for LVM-based gait recognition, termed BiggerGait. Comprehensive evaluations on CCPG, CAISA-B*, SUSTech1K, and CCGR_MINI validate the superiority of BiggerGait across both within- and cross-domain tasks, establishing it as a simple yet practical baseline for gait representation learning. All the models and code will be publicly available.
Problem

Research questions and friction points this paper is trying to address.

Investigates layer-wise LVM representations for gait recognition
Proposes BiggerGait as a universal baseline method
Validates superiority in cross-domain gait tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leveraging layer-wise LVM representations for gait recognition
Integrating intermediate layers for complementary task properties
Simple universal baseline without complex gait priors
🔎 Similar Papers
No similar papers found.
D
Dingqing Ye
Department of Computer Science and Engineering, Southern University of Science and Technology
C
Chao Fan
School of Artificial Intelligence, Shenzhen University
Z
Zhanbo Huang
Michigan State University
Chengwen Luo
Chengwen Luo
Shenzhen University
J
Jianqiang Li
School of Artificial Intelligence, Shenzhen University
Shiqi Yu
Shiqi Yu
Department of Computer Science and Engineering, Southern University of Science and Technology
gait recognitionbiometricsface detectioncomputer vision
X
Xiaoming Liu
Michigan State University