🤖 AI Summary
This work addresses the fundamental trade-off in long-term time series forecasting between parallel efficiency and temporal consistency: direct multi-step (DMS) approaches often lack sequential coherence, while iterative methods (IMS) suffer from error accumulation and slow inference. To overcome these limitations, we propose BTTF, a novel two-stage framework featuring lookahead enhancement and parallel self-refinement. BTTF leverages the model’s own predictions as augmentation signals and employs a lightweight ensemble to effectively refine a base forecaster. Notably, our approach requires no complex architectural modifications and is compatible with both linear and lightweight models. Extensive experiments demonstrate that BTTF achieves up to a 58% accuracy improvement across diverse long-horizon forecasting tasks, delivering consistent gains even when the base model is undertrained—thereby substantially transcending the inherent constraints of both DMS and IMS paradigms.
📝 Abstract
Long-term time series forecasting (LTSF) remains challenging due to the trade-off between parallel efficiency and sequential modeling of temporal coherence. Direct multi-step forecasting (DMS) methods enable fast, parallel prediction of all future horizons but often lose temporal consistency across steps, while iterative multi-step forecasting (IMS) preserves temporal dependencies at the cost of error accumulation and slow inference. To bridge this gap, we propose Back to the Future (BTTF), a simple yet effective framework that enhances forecasting stability through look-ahead augmentation and self-corrective refinement. Rather than relying on complex model architectures, BTTF revisits the fundamental forecasting process and refines a base model by ensembling the second-stage models augmented with their initial predictions. Despite its simplicity, our approach consistently improves long-horizon accuracy and mitigates the instability of linear forecasting models, achieving accuracy gains of up to 58% and demonstrating stable improvements even when the first-stage model is trained under suboptimal conditions. These results suggest that leveraging model-generated forecasts as augmentation can be a simple yet powerful way to enhance long-term prediction, even without complex architectures.