TimeDiT: General-purpose Diffusion Transformers for Time Series Foundation Model

📅 2024-09-03
🏛️ arXiv.org
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses key challenges in time-series modeling—including missing data imputation, multi-resolution analysis, uncertainty quantification, and integration of physical constraints—by proposing TimeDiT, the first general-purpose foundation model for time series. Methodologically, TimeDiT unifies Transformer-based representation learning with diffusion-based generative modeling, featuring a theory-driven unified masking schedule and knowledge-aware denoising process that enables consistent zero-shot and fine-tuned inference across diverse tasks. It introduces the first provably sound, fine-tuning-free model editing strategy, allowing dynamic injection of external domain knowledge during sampling. Extensive experiments demonstrate that TimeDiT achieves state-of-the-art performance across five core tasks: forecasting, imputation, multi-resolution modeling, anomaly detection, and generation. This is the first empirical validation of a prototype foundation model for time series, establishing its feasibility and broad generalization capability.

Technology Category

Application Category

📝 Abstract
Foundation models, particularly Large Language Models (LLMs), have revolutionized text and video processing, yet time series data presents distinct challenges for such approaches due to domain-specific features such as missing values, multi-resolution characteristics, etc. Furthermore, the de-facto autoregressive transformers tend to learn deterministic temporal dependencies within pre-trained data while overlooking inherent uncertainties and lacking integration of physical constraints. In this paper, we introduce TimeDiT, a diffusion transformer model that synergistically combines transformer-based temporal dependency learning with diffusion-based probabilistic sampling. TimeDiT employs a unified masking mechanism to harmonize the training and inference process across diverse tasks while introducing a theoretically grounded, finetuning-free model editing strategy that enables flexible integration of external knowledge during sampling. Acknowledging the challenges of unifying multiple downstream tasks under a single model, our systematic evaluation demonstrates TimeDiT's effectiveness both in fundamental tasks, i.e., forecasting and imputation, through zero-shot/fine-tuning; and in domain tasks, i.e., multi-resolution forecasting, anomaly detection, and data generation, establishing it as a extit{proto-foundation model} that bridges the gap between general-purpose and domain-specific models.
Problem

Research questions and friction points this paper is trying to address.

Time series data challenges
Autoregressive transformers limitations
Unified model for diverse tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Diffusion transformer model
Unified masking mechanism
Finetuning-free model editing
🔎 Similar Papers
No similar papers found.
Defu Cao
Defu Cao
Peking University; MBZUAI; University of Southern California; Caltech
Time SeriesFoundation ModelMachine LearningCausal InferenceLLM
W
Wen Ye
Department of Computer Science, University of Southern California
Y
Yizhou Zhang
Department of Computer Science, University of Southern California
Y
Yan Liu
Department of Computer Science, University of Southern California