T1: One-to-One Channel-Head Binding for Multivariate Time-Series Imputation

📅 2026-02-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of imputing multivariate time series under high missing rates and complex missing patterns, where existing methods suffer from degraded temporal features that hinder effective cross-variable information transfer and lead to large reconstruction errors. To overcome these limitations, the authors propose the T1 model, which employs a CNN-Transformer hybrid architecture featuring an innovative one-to-one binding mechanism between channels and attention heads. This design jointly optimizes temporal feature extraction and cross-variable dependency modeling, while adaptive attention weighting suppresses pathways corrupted by missing data. Evaluated on 11 benchmark datasets, T1 achieves state-of-the-art performance, reducing average MSE by 46% compared to the next-best method, maintains robustness even at 70% missingness, and demonstrates zero-shot generalization to unseen missing patterns without retraining.

Technology Category

Application Category

📝 Abstract
Imputing missing values in multivariate time series remains challenging, especially under diverse missing patterns and heavy missingness. Existing methods suffer from suboptimal performance as corrupted temporal features hinder effective cross-variable information transfer, amplifying reconstruction errors. Robust imputation requires both extracting temporal patterns from sparse observations within each variable and selectively transferring information across variables--yet current approaches excel at one while compromising the other. We introduce T1 (Time series imputation with 1-to-1 channel-head binding), a CNN-Transformer hybrid architecture that achieves robust imputation through Channel-Head Binding--a mechanism creating one-to-one correspondence between CNN channels and attention heads. This design enables selective information transfer: when missingness corrupts certain temporal patterns, their corresponding attention pathways adaptively down-weight based on remaining observable patterns while preserving reliable cross-variable connections through unaffected channels. Experiments on 11 benchmark datasets demonstrate that T1 achieves state-of-the-art performance, reducing MSE by 46% on average compared to the second-best baseline, with particularly strong gains under extreme sparsity (70% missing ratio). The model generalizes to unseen missing patterns without retraining and uses a consistent hyperparameter configuration across all datasets. The code is available at https://github.com/Oppenheimerdinger/T1.
Problem

Research questions and friction points this paper is trying to address.

multivariate time-series imputation
missing data
temporal patterns
cross-variable information transfer
missingness
Innovation

Methods, ideas, or system contributions that make the work stand out.

Channel-Head Binding
multivariate time-series imputation
CNN-Transformer hybrid
selective cross-variable transfer
robust imputation under sparsity
🔎 Similar Papers
No similar papers found.
D
Dongik Park
Graduate School of Data Science, Seoul National University
Hyunwoo Ryu
Hyunwoo Ryu
MIT
Artificial IntelligenceRobotics
S
Suahn Bae
Graduate School of Data Science, Seoul National University
K
Keondo Park
Graduate School of Data Science, Seoul National University
Hyung-Sin Kim
Hyung-Sin Kim
Seoul National University, Data Science
On-device AIMachine learningComputer visionInternet of Things