Identity Clue Refinement and Enhancement for Visible-Infrared Person Re-Identification

📅 2025-12-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Visible-light–infrared person re-identification (VI-ReID) suffers from significant modality discrepancy, yet existing methods overemphasize modality-invariant features while neglecting the discriminative value of modality-specific identity cues. To address this, we propose an identity-cue-driven cross-modal learning framework. First, a multi-perceptive feature refinement module explicitly models subtle, often-overlooked modality-specific attributes. Second, a semantic distillation cascaded enhancement mechanism jointly optimizes both modality-invariant and modality-specific representations. Third, an identity-cue-guided loss is introduced to enhance discriminability and diversity in the feature space. Our method integrates shallow-feature aggregation with knowledge distillation strategies. Extensive experiments on standard benchmarks—including SYSU-MM01 and RegDB—demonstrate substantial improvements over state-of-the-art methods, validating the critical role of modality-specific identity knowledge in advancing VI-ReID performance.

Technology Category

Application Category

📝 Abstract
Visible-Infrared Person Re-Identification (VI-ReID) is a challenging cross-modal matching task due to significant modality discrepancies. While current methods mainly focus on learning modality-invariant features through unified embedding spaces, they often focus solely on the common discriminative semantics across modalities while disregarding the critical role of modality-specific identity-aware knowledge in discriminative feature learning. To bridge this gap, we propose a novel Identity Clue Refinement and Enhancement (ICRE) network to mine and utilize the implicit discriminative knowledge inherent in modality-specific attributes. Initially, we design a Multi-Perception Feature Refinement (MPFR) module that aggregates shallow features from shared branches, aiming to capture modality-specific attributes that are easily overlooked. Then, we propose a Semantic Distillation Cascade Enhancement (SDCE) module, which distills identity-aware knowledge from the aggregated shallow features and guide the learning of modality-invariant features. Finally, an Identity Clues Guided (ICG) Loss is proposed to alleviate the modality discrepancies within the enhanced features and promote the learning of a diverse representation space. Extensive experiments across multiple public datasets clearly show that our proposed ICRE outperforms existing SOTA methods.
Problem

Research questions and friction points this paper is trying to address.

Mines modality-specific identity-aware knowledge for cross-modal matching
Refines shallow features to capture overlooked modality-specific attributes
Reduces modality discrepancies to enhance discriminative feature learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-Perception Feature Refinement captures modality-specific attributes
Semantic Distillation Cascade Enhancement distills identity-aware knowledge
Identity Clues Guided Loss alleviates modality discrepancies in features
🔎 Similar Papers
No similar papers found.
G
Guoqing Zhang
School of Computer Science, Nanjing University of Information Science and Technology, and also with the Key Laboratory of Social Computing and Cognitive Intelligence (Dalian University of Technology), Ministry of Education, Dalian, China
Zhun Wang
Zhun Wang
Graduate Student, UC Berkeley
H
Hairui Wang
School of Computer Science, Nanjing University of Information Science and Technology, Nanjing, 210044, China
Z
Zhonglin Ye
State Key Laboratory of Tibetan Intelligent Information Processing and Application, Qinghai Normal University, Xining, 810008, China
Yuhui Zheng
Yuhui Zheng
Full Professor with school of Computer and Software, NUIST
Computer Vision、Multimedia Forensics、Digital Watermarking