Baguan-TS: A Sequence-Native In-Context Learning Model for Time Series Forecasting with Covariates

📅 2026-03-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes Baguan-TS, the first native sequence-based in-context learning framework for covariate time series forecasting. Existing approaches either rely on handcrafted features or lack gradient-free rapid adaptation at inference time, struggling to balance end-to-end modeling with efficient generalization. Baguan-TS addresses these limitations through a unified three-dimensional Transformer architecture that jointly models temporal, variate, and contextual dimensions. It further incorporates target-space retrieval calibration and a context overfitting mechanism to mitigate training instability and excessive output smoothing. Evaluated across multiple public benchmarks and real-world energy datasets, Baguan-TS consistently achieves state-of-the-art performance and robustness in both point and probabilistic forecasting tasks.

Technology Category

Application Category

📝 Abstract
Transformers enable in-context learning (ICL) for rapid, gradient-free adaptation in time series forecasting, yet most ICL-style approaches rely on tabularized, hand-crafted features, while end-to-end sequence models lack inference-time adaptation. We bridge this gap with a unified framework, Baguan-TS, which integrates the raw-sequence representation learning with ICL, instantiated by a 3D Transformer that attends jointly over temporal, variable, and context axes. To make this high-capacity model practical, we tackle two key hurdles: (i) calibration and training stability, improved with a feature-agnostic, target-space retrieval-based local calibration; and (ii) output oversmoothing, mitigated via context-overfitting strategy. On public benchmark with covariates, Baguan-TS consistently outperforms established baselines, achieving the highest win rate and significant reductions in both point and probabilistic forecasting metrics. Further evaluations across diverse real-world energy datasets demonstrate its robustness, yielding substantial improvements.
Problem

Research questions and friction points this paper is trying to address.

in-context learning
time series forecasting
sequence modeling
covariates
adaptation gap
Innovation

Methods, ideas, or system contributions that make the work stand out.

In-Context Learning
Time Series Forecasting
3D Transformer
Sequence-Native Representation
Context-Overfitting
🔎 Similar Papers
No similar papers found.
Linxiao Yang
Linxiao Yang
Alibaba Group
XAI
X
Xue Jiang
1DAMO Academy, Alibaba Group, Hangzhou, China.
Gezheng Xu
Gezheng Xu
University of Western Ontario
Deep LearningMachine Learning
Tian Zhou
Tian Zhou
Damo-Alibaba Group
Time Series AnalysisMachine learningSpatial Temporal ForecastingWeather Forecasting
Min Yang
Min Yang
Bytedance
Vision Language ModelComputer VisionVideo Understanding
Z
ZhaoYang Zhu
1DAMO Academy, Alibaba Group, Hangzhou, China.
L
Linyuan Geng
1DAMO Academy, Alibaba Group, Hangzhou, China.
Z
Zhipeng Zeng
1DAMO Academy, Alibaba Group, Hangzhou, China.
Q
Qiming Chen
1DAMO Academy, Alibaba Group, Hangzhou, China.
X
Xinyue Gu
1DAMO Academy, Alibaba Group, Hangzhou, China.
Rong Jin
Rong Jin
Alibaba Group
machine learningoptimization
Liang Sun
Liang Sun
DAMO Academy, Alibaba Group
AI4Weathertime seriesanomaly detectionforecastingXAI