TimeMKG: Knowledge-Infused Causal Reasoning for Multivariate Time Series Modeling

📅 2025-08-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional multivariate time series modeling often neglects the semantic information embedded in variable names and textual descriptions, resulting in limited model interpretability and poor generalization. To address this, we propose a semantics-enhanced dual-modality modeling framework. First, a large language model is employed to parse variable-level textual descriptions and construct a structured multivariate knowledge graph. Second, a dual-modality encoder is designed to jointly model semantic prompts and numerical time-series patterns via cross-modal attention, augmented with causal priors to support causal reasoning. This work represents the first systematic integration of fine-grained, variable-level semantic knowledge into the time series modeling pipeline. Extensive experiments on multiple benchmark datasets demonstrate significant improvements in both forecasting and classification performance, alongside enhanced model interpretability and cross-domain generalization capability.

Technology Category

Application Category

📝 Abstract
Multivariate time series data typically comprises two distinct modalities: variable semantics and sampled numerical observations. Traditional time series models treat variables as anonymous statistical signals, overlooking the rich semantic information embedded in variable names and data descriptions. However, these textual descriptors often encode critical domain knowledge that is essential for robust and interpretable modeling. Here we present TimeMKG, a multimodal causal reasoning framework that elevates time series modeling from low-level signal processing to knowledge informed inference. TimeMKG employs large language models to interpret variable semantics and constructs structured Multivariate Knowledge Graphs that capture inter-variable relationships. A dual-modality encoder separately models the semantic prompts, generated from knowledge graph triplets, and the statistical patterns from historical time series. Cross-modality attention aligns and fuses these representations at the variable level, injecting causal priors into downstream tasks such as forecasting and classification, providing explicit and interpretable priors to guide model reasoning. The experiment in diverse datasets demonstrates that incorporating variable-level knowledge significantly improves both predictive performance and generalization.
Problem

Research questions and friction points this paper is trying to address.

Modeling multivariate time series without utilizing semantic variable information
Overlooking domain knowledge embedded in variable names and descriptions
Lack of interpretable causal reasoning in traditional time series approaches
Innovation

Methods, ideas, or system contributions that make the work stand out.

Knowledge-infused causal reasoning framework
Multivariate Knowledge Graphs from variable semantics
Cross-modality attention fuses semantic and statistical patterns
Y
Yifei Sun
Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai, 200237, P. R. China
J
Junming Liu
Shanghai Artificial Intelligence Laboratory
D
Ding Wang
Shanghai Artificial Intelligence Laboratory
Yirong Chen
Yirong Chen
Stanford University
Xuefeng Yan
Xuefeng Yan
Molecular Imaging Branch/National Institute of Mental Health/National Institutes of Health
Molecular imaging