MTS-DMAE: Dual-Masked Autoencoder for Unsupervised Multivariate Time Series Representation Learning

📅 2025-09-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Unsupervised multivariate time series representation learning suffers from weak temporal consistency and insufficient semantic information in learned representations. To address this, we propose a dual-masked autoencoder framework: masks are applied jointly at both the input layer and latent space, enabling co-optimization of masked value reconstruction and latent representation prediction. A momentum-updated teacher encoder is introduced, coupled with feature-level alignment constraints, to jointly optimize observable attributes and latent structural patterns. This design significantly enhances temporal robustness and semantic discriminability of the learned representations. Extensive experiments demonstrate that our method consistently outperforms state-of-the-art unsupervised baselines across diverse downstream tasks—including classification, regression, and forecasting—on multiple benchmark datasets. Moreover, the learned representations exhibit strong transferability and generalization capability.

Technology Category

Application Category

📝 Abstract
Unsupervised multivariate time series (MTS) representation learning aims to extract compact and informative representations from raw sequences without relying on labels, enabling efficient transfer to diverse downstream tasks. In this paper, we propose Dual-Masked Autoencoder (DMAE), a novel masked time-series modeling framework for unsupervised MTS representation learning. DMAE formulates two complementary pretext tasks: (1) reconstructing masked values based on visible attributes, and (2) estimating latent representations of masked features, guided by a teacher encoder. To further improve representation quality, we introduce a feature-level alignment constraint that encourages the predicted latent representations to align with the teacher's outputs. By jointly optimizing these objectives, DMAE learns temporally coherent and semantically rich representations. Comprehensive evaluations across classification, regression, and forecasting tasks demonstrate that our approach achieves consistent and superior performance over competitive baselines.
Problem

Research questions and friction points this paper is trying to address.

Unsupervised representation learning for multivariate time series
Reconstructing masked values and estimating latent representations
Improving representation quality with feature-level alignment constraint
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-masked autoencoder for unsupervised representation learning
Reconstructing masked values and estimating latent representations
Feature-level alignment constraint with teacher encoder guidance
🔎 Similar Papers
No similar papers found.
Y
Yi Xu
Northeastern University, Boston, USA
Yitian Zhang
Yitian Zhang
Northeastern University
computer vision
Y
Yun Fu
Northeastern University, Boston, USA