SPADE-S: A Sparsity-Robust Foundational Forecaster

📅 2025-07-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address systematic prediction bias in low-magnitude and sparse time series—arising from magnitude heterogeneity and data sparsity—this paper proposes a robust time-series forecasting architecture. Methodologically, it introduces: (1) a debiased loss function to mitigate gradient imbalance induced by amplitude disparities; (2) a dynamic sparsity-aware sampling strategy to enhance training stability on sparse sequences; and (3) a temporal-adaptive encoder to strengthen multi-scale pattern representation. This is the first work to systematically tackle performance degradation of deep forecasting models on heterogeneous time series. Evaluated on three large-scale e-commerce datasets containing 3 million to 700 million series, the method achieves P90 accuracy gains of 2.21%–6.58%, a P50 improvement of 1.95%, and up to 15% single-point accuracy gain over state-of-the-art baselines.

Technology Category

Application Category

📝 Abstract
Despite significant advancements in time series forecasting, accurate modeling of time series with strong heterogeneity in magnitude and/or sparsity patterns remains challenging for state-of-the-art deep learning architectures. We identify several factors that lead existing models to systematically underperform on low-magnitude and sparse time series, including loss functions with implicit biases toward high-magnitude series, training-time sampling methods, and limitations of time series encoding methods. SPADE-S is a robust forecasting architecture that significantly reduces magnitude- and sparsity-based systematic biases and improves overall prediction accuracy. Empirical results demonstrate that SPADE-S outperforms existing state-of-the-art approaches across a diverse set of use cases in demand forecasting. In particular, we show that, depending on the quantile forecast and magnitude of the series, SPADE-S can improve forecast accuracy by up to 15%. This results in P90 overall forecast accuracy gains of 2.21%, 6.58%, and 4.28%, and P50 forecast accuracy gains of 0.92%, 0.77%, and 1.95%, respectively, for each of three distinct datasets, ranging from 3 million to 700 million series, from a large online retailer.
Problem

Research questions and friction points this paper is trying to address.

Addresses biases in forecasting low-magnitude sparse time series
Improves accuracy for heterogeneous magnitude and sparsity patterns
Enhances demand forecasting across diverse large-scale datasets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Robust architecture for heterogeneous time series
Reduces magnitude and sparsity biases
Improves forecast accuracy up to 15%
🔎 Similar Papers
No similar papers found.
M
Malcolm Wolff
Amazon SCOT Forecasting. New York, New York. USA.
M
Matthew Li
Amazon SCOT Forecasting. New York, New York. USA.
R
Ravi Kiran Selvam
Amazon SCOT Forecasting. New York, New York. USA.
H
Hanjing Zhu
Amazon SCOT Forecasting. New York, New York. USA.
Kin G. Olivares
Kin G. Olivares
Carnegie Mellon University
machine learningstatisticsforecasting
R
Ruijun Ma
Amazon SCOT Forecasting. New York, New York. USA.
A
Abhinav Katoch
Amazon SCOT Forecasting. New York, New York. USA.
S
Shankar Ramasubramanian
Amazon SCOT Forecasting. New York, New York. USA.
M
Mengfei Cao
Amazon SCOT Forecasting. New York, New York. USA.
R
Roberto Bandarra
Amazon SCOT Forecasting. New York, New York. USA.
R
Rahul Gopalsamy
Amazon SCOT Forecasting. New York, New York. USA.
S
Stefania La Vattiata
S
Sitan Yang
Amazon SCOT Forecasting. New York, New York. USA.
M
Michael M. Mahoney
Amazon SCOT Forecasting. New York, New York. USA.