Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Spatiotemporal-Linear: Towards Universal Multivariate Time Series Forecasting (2312.14869v1)

Published 22 Dec 2023 in cs.LG

Abstract: Within the field of complicated multivariate time series forecasting (TSF), popular techniques frequently rely on intricate deep learning architectures, ranging from transformer-based designs to recurrent neural networks. However, recent findings suggest that simple Linear models can surpass sophisticated constructs on diverse datasets. These models directly map observation to multiple future time steps, thereby minimizing error accumulation in iterative multi-step prediction. Yet, these models fail to incorporate spatial and temporal information within the data, which is critical for capturing patterns and dependencies that drive insightful predictions. This oversight often leads to performance bottlenecks, especially under specific sequence lengths and dataset conditions, preventing their universal application. In response, we introduce the SpatioTemporal-Linear (STL) framework. STL seamlessly integrates time-embedded and spatially-informed bypasses to augment the Linear-based architecture. These extra routes offer a more robust and refined regression to the data, particularly when the amount of observation is limited and the capacity of simple linear layers to capture dependencies declines. Empirical evidence highlights STL's prowess, outpacing both Linear and Transformer benchmarks across varied observation and prediction durations and datasets. Such robustness accentuates its suitability across a spectrum of applications, including but not limited to, traffic trajectory and rare disease progression forecasting. Through this discourse, we not only validate the STL's distinctive capacities to become a more general paradigm in multivariate time-series prediction using deep-learning techniques but also stress the need to tackle data-scarce prediction scenarios for universal application. Code will be made available.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Application of neural networks to an emerging financial market: forecasting and trading the taiwan stock index. Computers & Operations Research, 30(6):901–923, 2003.
  2. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pages 11106–11115, 2021.
  3. Network traffic classifier with convolutional and recurrent neural networks for internet of things. IEEE access, 5:18042–18050, 2017.
  4. Joint attention in autonomous driving (jaad). arXiv preprint arXiv:1609.04741, 2016.
  5. An integrated neural network model for pm10 forecasting. Atmospheric environment, 40(16):2845–2851, 2006.
  6. Statistics for spatio-temporal data. Hoboken, N.J. Wiley, 2011.
  7. Larry R Medsker and LC Jain. Recurrent neural networks. Design and Applications, 5(64-67):2, 2001.
  8. Temporal convolutional networks for action segmentation and detection. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 156–165, 2017.
  9. A survey on deep learning for time-series forecasting. Machine learning and big data analytics paradigms: analysis, applications and challenges, pages 365–392, 2021.
  10. Transformers in time series: A survey. arXiv preprint arXiv:2202.07125, 2022.
  11. Long-range transformers for dynamic spatiotemporal forecasting. arXiv preprint arXiv:2109.12218, 2021.
  12. Temporal fusion transformers for interpretable multi-horizon time series forecasting. International Journal of Forecasting, 37(4):1748–1764, 2021.
  13. Are transformers effective for time series forecasting? In Proceedings of the AAAI conference on artificial intelligence, volume 37, pages 11121–11128, 2023.
  14. Charles M Jones. What do we know about high-frequency trading? Columbia Business School Research Paper, (13-11), 2013.
  15. Pie: A large-scale dataset and models for pedestrian intention estimation and trajectory prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6262–6271, 2019.
  16. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  17. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  18. A survey on vision transformer. IEEE transactions on pattern analysis and machine intelligence, 45(1):87–110, 2022.
  19. Time-series forecasting with deep learning: a survey. Philosophical Transactions of the Royal Society A, 379(2194):20200209, 2021.
  20. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014.
  21. Learning to forget: Continual prediction with lstm. Neural computation, 12(10):2451–2471, 2000.
  22. Brian K Nelson. Time series analysis using autoregressive integrated moving average (arima) models. Academic emergency medicine, 5(7):739–744, 1998.
  23. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In International conference on learning representations, 2021.
  24. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International Conference on Machine Learning, pages 27268–27286. PMLR, 2022.
  25. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, 34:22419–22430, 2021.
  26. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Advances in neural information processing systems, 32, 2019.
  27. Spatial-temporal transformer networks for traffic flow forecasting. arXiv preprint arXiv:2001.02908, 2020.
  28. Spatiotemporal transformer neural network for time-series forecasting. Entropy, 24(11):1651, 2022.
  29. Poseformerv2: Exploring frequency domain for efficient and robust 3d human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8877–8886, 2023.
  30. Part aware contrastive learning for self-supervised action recognition. In International Joint Conference on Artificial Intelligence, 2023.
  31. Smart video surveillance: exploring the concept of multiscale spatiotemporal tracking. IEEE signal processing magazine, 22(2):38–51, 2005.
  32. Spatiotemporal graph neural network for performance prediction of photovoltaic power systems. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 15323–15330, 2021.
  33. Parallel spatio-temporal attention-based tcn for multivariate time series prediction. Neural Computing and Applications, 35(18):13109–13118, 2023.
  34. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
Citations (1)

Summary

We haven't generated a summary for this paper yet.