Dynamic Dictionary Learning for Remote Sensing Image Segmentation

📅 2025-03-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Remote sensing image segmentation faces challenges in distinguishing morphologically similar categories and adapting to diverse scenes—particularly in fine-grained tasks such as cloud thickness classification, where existing methods struggle to dynamically model context-aware semantic embeddings. To address this, we propose a dynamic dictionary learning framework: (1) multi-stage alternating cross-attention iteratively refines image features and class-aware semantic embeddings; (2) a novel differentiable dynamic dictionary construction mechanism, jointly optimized with dictionary-level contrastive constraints, explicitly balances intra-class heterogeneity and inter-class homogeneity—overcoming limitations of implicit representation learning. Extensive experiments demonstrate state-of-the-art performance on the LoveDA and UAVid online test sets. Crucially, our method achieves consistent improvements across both coarse- and fine-grained benchmarks, including cloud thickness segmentation.

Technology Category

Application Category

📝 Abstract
Remote sensing image segmentation faces persistent challenges in distinguishing morphologically similar categories and adapting to diverse scene variations. While existing methods rely on implicit representation learning paradigms, they often fail to dynamically adjust semantic embeddings according to contextual cues, leading to suboptimal performance in fine-grained scenarios such as cloud thickness differentiation. This work introduces a dynamic dictionary learning framework that explicitly models class ID embeddings through iterative refinement. The core contribution lies in a novel dictionary construction mechanism, where class-aware semantic embeddings are progressively updated via multi-stage alternating cross-attention querying between image features and dictionary embeddings. This process enables adaptive representation learning tailored to input-specific characteristics, effectively resolving ambiguities in intra-class heterogeneity and inter-class homogeneity. To further enhance discriminability, a contrastive constraint is applied to the dictionary space, ensuring compact intra-class distributions while maximizing inter-class separability. Extensive experiments across both coarse- and fine-grained datasets demonstrate consistent improvements over state-of-the-art methods, particularly in two online test benchmarks (LoveDA and UAVid). Code is available at https://anonymous.4open.science/r/D2LS-8267/.
Problem

Research questions and friction points this paper is trying to address.

Dynamic dictionary learning for remote sensing image segmentation.
Resolving ambiguities in intra-class heterogeneity and inter-class homogeneity.
Improving performance in fine-grained scenarios like cloud thickness differentiation.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic dictionary learning for image segmentation
Multi-stage cross-attention for semantic embedding updates
Contrastive constraints enhance class separability
🔎 Similar Papers
No similar papers found.
X
Xuechao Zou
School of Computer Science and Technology, Beijing Jiaotong University, Beijing, China
Y
Yue Li
Department of Computer Technology and Applications, Qinghai University, Xining, China
S
Shun Zhang
Department of Computer Technology and Applications, Qinghai University, Xining, China
K
Kai Li
Department of Computer Science and Technology, Tsinghua University, Beijing, China
Shiying Wang
Shiying Wang
Yale University
GeneticsNeuroimaging.
P
Pin Tao
Department of Computer Technology and Applications, Qinghai University, Xining, China; Department of Computer Science and Technology, Tsinghua University, Beijing, China
J
Junliang Xing
Department of Computer Science and Technology, Tsinghua University, Beijing, China
Congyan Lang
Congyan Lang
Beijing Jiaotong University
computer vision