T-LLM: Teaching Large Language Models to Forecast Time Series via Temporal Distillation

📅 2026-02-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) struggle with time series forecasting due to their lack of explicit temporal modeling capabilities. This work proposes a temporal distillation framework that, for the first time, injects the predictive prowess of a lightweight teacher model into a general-purpose LLM through behavioral distillation. The teacher model integrates trend modeling and frequency-domain analysis to provide structured supervisory signals during training. Once trained, the LLM alone can perform end-to-end forecasting without requiring any additional inference modules. Evaluated across multiple benchmarks and infectious disease prediction tasks, the proposed approach consistently outperforms existing LLM-based forecasting methods under full-data, few-shot, and zero-shot settings.

Technology Category

Application Category

📝 Abstract
Time series forecasting plays a critical role in decision-making across many real-world applications. Unlike data in vision and language domains, time series data is inherently tied to the evolution of underlying processes and can only accumulate as real-world time progresses, limiting the effectiveness of scale-driven pretraining alone. This time-bound constraint poses a challenge for enabling large language models (LLMs) to acquire forecasting capability, as existing approaches primarily rely on representation-level alignment or inference-time temporal modules rather than explicitly teaching forecasting behavior to the LLM. We propose T-LLM, a temporal distillation framework that equips general-purpose LLMs with time series forecasting capability by transferring predictive behavior from a lightweight temporal teacher during training. The teacher combines trend modeling and frequency-domain analysis to provide structured temporal supervision, and is removed entirely at inference, leaving the LLM as the sole forecasting model. Experiments on benchmark datasets and infectious disease forecasting tasks demonstrate that T-LLM consistently outperforms existing LLM-based forecasting methods under full-shot, few-shot, and zero-shot settings, while enabling a simple and efficient deployment pipeline.
Problem

Research questions and friction points this paper is trying to address.

time series forecasting
large language models
temporal distillation
forecasting capability
time-bound data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Temporal Distillation
Time Series Forecasting
Large Language Models
Teacher-Student Framework
Frequency-Domain Analysis
🔎 Similar Papers
No similar papers found.
S
Suhan Guo
State Key Laboratory for Novel Software Technology, Nanjing University; School of Artificial Intelligence, Nanjing University
B
Bingxu Wang
State Key Laboratory for Novel Software Technology, Nanjing University; School of Artificial Intelligence, Nanjing University
S
Shaodan Zhang
State Key Laboratory for Novel Software Technology, Nanjing University; School of Artificial Intelligence, Nanjing University
Furao Shen
Furao Shen
Department of Computer Science & Technology, Nanjing University
Neural NetworksRobotic Intelligence