π€ AI Summary
Existing cross-spectral patch matching methods overemphasize inter-patch relationships while neglecting intrinsic feature modeling of individual patches. To address this, we propose a unified relational representation learning frameworkβthe first to jointly model both intrinsic patch features and inter-patch relationships. Specifically, we design a joint learning mechanism integrating intrinsic feature auto-encoding representation with deep feature interaction; introduce a lightweight Multi-dimensional Global-to-Local Attention (MGLA) module to enhance multi-scale contextual awareness; and develop an Attention-driven Lightweight Feature Extraction (ALFE) network coupled with a Multi-Loss Post-Pruning (MLPP) optimization strategy. Evaluated on multiple public benchmarks, our framework achieves state-of-the-art performance, significantly improving matching accuracy without increasing model parameter count or inference latency.
π Abstract
Recently, feature relation learning has drawn widespread attention in cross-spectral image patch matching. However, existing related research focuses on extracting diverse relations between image patch features and ignores sufficient intrinsic feature representations of individual image patches. Therefore, we propose an innovative relational representation learning idea that simultaneously focuses on sufficiently mining the intrinsic features of individual image patches and the relations between image patch features. Based on this, we construct a Relational Representation Learning Network (RRL-Net). Specifically, we innovatively construct an autoencoder to fully characterize the individual intrinsic features, and introduce a feature interaction learning (FIL) module to extract deep-level feature relations. To further fully mine individual intrinsic features, a lightweight multi-dimensional global-to-local attention (MGLA) module is constructed to enhance the global feature extraction of individual image patches and capture local dependencies within global features. By combining the MGLA module, we further explore the feature extraction network and construct an attention-based lightweight feature extraction (ALFE) network. In addition, we propose a multi-loss post-pruning (MLPP) optimization strategy, which greatly promotes network optimization while avoiding increases in parameters and inference time. Extensive experiments demonstrate that our RRL-Net achieves state-of-the-art (SOTA) performance on multiple public datasets. Our code are available at https://github.com/YuChuang1205/RRL-Net.