🤖 AI Summary
Shallow models for multivariate spatiotemporal (especially 4D) clustering suffer from ignoring clustering errors, imbalanced local–global feature representation, and insufficient modeling of long-range dependencies and spatial positions.
Method: We propose a Deep Adversarial Attention Subspace Clustering (DAASC) framework, featuring a U-Net–style deep generative-discriminative architecture. It integrates TimeDistributed ConvLSTM2D with graph attention Transformers to jointly capture local spatiotemporal structures, global dependencies, and long-range correlations. A self-expressive layer coupled with adversarial training is introduced to jointly optimize subspace representations.
Contribution/Results: Evaluated on three real-world multivariate spatiotemporal datasets, DAASC consistently outperforms state-of-the-art methods, achieving average improvements of 5.2–9.8% in clustering accuracy (ACC/NMI/ARI). This work provides the first systematic empirical validation of the effectiveness and generalizability of deep adversarial attention mechanisms for 4D spatiotemporal subspace clustering.
📝 Abstract
Deep subspace clustering models are vital for applications such as snowmelt detection, sea ice tracking, crop health monitoring, infectious disease modeling, network load prediction, and land-use planning, where multivariate spatiotemporal data exhibit complex temporal dependencies and reside on multiple nonlinear manifolds beyond the capability of traditional clustering methods. These models project data into a latent space where samples lie in linear subspaces and exploit the self-expressiveness property to uncover intrinsic relationships. Despite their success, existing methods face major limitations: they use shallow autoencoders that ignore clustering errors, emphasize global features while neglecting local structure, fail to model long-range dependencies and positional information, and are rarely applied to 4D spatiotemporal data. To address these issues, we propose A-DATSC (Attention-Guided Deep Adversarial Temporal Subspace Clustering), a model combining a deep subspace clustering generator and a quality-verifying discriminator. The generator, inspired by U-Net, preserves spatial and temporal integrity through stacked TimeDistributed ConvLSTM2D layers, reducing parameters and enhancing generalization. A graph attention transformer based self-expressive network captures local spatial relationships, global dependencies, and both short- and long-range correlations. Experiments on three real-world multivariate spatiotemporal datasets show that A-DATSC achieves substantially superior clustering performance compared to state-of-the-art deep subspace clustering models.