🤖 AI Summary
To address the low adoption and performance instability of ensemble methods—particularly stacking—in time-series forecasting, this paper proposes a multi-layer stacking framework that hierarchically fuses heterogeneous base models with meta-learners, mitigating the task-transfer volatility inherent in single-stack ensembles. The method systematically integrates 33 ensemble models, including a newly designed nonlinear combination mechanism. We conduct large-scale empirical evaluation across 50 real-world time-series datasets. Results demonstrate that the framework consistently outperforms conventional linear-weighted ensembling and other state-of-the-art approaches across diverse forecasting tasks, delivering stable and statistically significant accuracy improvements. This work establishes a scalable, robust paradigm for time-series ensemble learning, advancing both methodological design and practical deployment of stacked ensembles.
📝 Abstract
Ensembling is a powerful technique for improving the accuracy of machine learning models, with methods like stacking achieving strong results in tabular tasks. In time series forecasting, however, ensemble methods remain underutilized, with simple linear combinations still considered state-of-the-art. In this paper, we systematically explore ensembling strategies for time series forecasting. We evaluate 33 ensemble models -- both existing and novel -- across 50 real-world datasets. Our results show that stacking consistently improves accuracy, though no single stacker performs best across all tasks. To address this, we propose a multi-layer stacking framework for time series forecasting, an approach that combines the strengths of different stacker models. We demonstrate that this method consistently provides superior accuracy across diverse forecasting scenarios. Our findings highlight the potential of stacking-based methods to improve AutoML systems for time series forecasting.