Unlocking the Power of LSTM for Long Term Time Series Forecasting

๐Ÿ“… 2024-08-19
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 3
โœจ Influential: 1
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the limitation of sLSTMโ€”its poor short-term memory retention, which hinders direct application to long-horizon time series forecasting (TSF)โ€”this paper proposes P-sLSTM, the first model to synergistically integrate sequence patching and channel independence. P-sLSTM enhances local pattern modeling via patch-based encoding, while leveraging channel-wise decoupling, exponential gating, and memory mixing to mitigate short-term memory degradation without compromising sLSTMโ€™s inherent capacity for long-range dependency capture. Theoretically interpretable, structurally lightweight, and parameter-efficient, P-sLSTM achieves state-of-the-art performance across multiple standard long-horizon TSF benchmarks, reducing average prediction error by 12.7% compared to prior methods. Moreover, it outperforms mainstream Transformer- and RNN-based variants in inference speed, demonstrating both accuracy and efficiency advantages.

Technology Category

Application Category

๐Ÿ“ Abstract
Traditional recurrent neural network architectures, such as long short-term memory neural networks (LSTM), have historically held a prominent role in time series forecasting (TSF) tasks. While the recently introduced sLSTM for Natural Language Processing (NLP) introduces exponential gating and memory mixing that are beneficial for long term sequential learning, its potential short memory issue is a barrier to applying sLSTM directly in TSF. To address this, we propose a simple yet efficient algorithm named P-sLSTM, which is built upon sLSTM by incorporating patching and channel independence. These modifications substantially enhance sLSTM's performance in TSF, achieving state-of-the-art results. Furthermore, we provide theoretical justifications for our design, and conduct extensive comparative and analytical experiments to fully validate the efficiency and superior performance of our model.
Problem

Research questions and friction points this paper is trying to address.

Enhance sLSTM for long-term TSF
Address short memory in sLSTM
Propose P-sLSTM with patching method
Innovation

Methods, ideas, or system contributions that make the work stand out.

P-sLSTM enhances sLSTM
Incorporates patching technique
Uses channel independence
๐Ÿ”Ž Similar Papers
No similar papers found.