Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 91 tok/s
Gemini 2.5 Pro 58 tok/s Pro
GPT-5 Medium 29 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 102 tok/s
GPT OSS 120B 462 tok/s Pro
Kimi K2 181 tok/s Pro
2000 character limit reached

Variational quantization for state space models (2404.11117v1)

Published 17 Apr 2024 in cs.LG

Abstract: Forecasting tasks using large datasets gathering thousands of heterogeneous time series is a crucial statistical problem in numerous sectors. The main challenge is to model a rich variety of time series, leverage any available external signals and provide sharp predictions with statistical guarantees. In this work, we propose a new forecasting model that combines discrete state space hidden Markov models with recent neural network architectures and training procedures inspired by vector quantized variational autoencoders. We introduce a variational discrete posterior distribution of the latent states given the observations and a two-stage training procedure to alternatively train the parameters of the latent states and of the emission distributions. By learning a collection of emission laws and temporarily activating them depending on the hidden process dynamics, the proposed method allows to explore large datasets and leverage available external signals. We assess the performance of the proposed method using several datasets and show that it outperforms other state-of-the-art solutions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. An Input Output HMM architecture. In Tesauro, G., Touretzky, D., and Leen, T., editors, Advances in Neural Information Processing Systems, volume 7. MIT Press.
  2. Time series analysis: forecasting and control. John Wiley & Sons.
  3. The fundamental theorem of exponential smoothing. Operations Research, 9(5):673–685.
  4. Nhits: Neural hierarchical interpolation for time series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 6989–6997.
  5. An introduction to sequential Monte Carlo. Springer.
  6. HERMES: Hybrid error-corrector model with inclusion of external signals for nonstationary fashion time series. preprint.
  7. Identifiability of discrete Input-output hidden Markov models with external signals. working paper or preprint.
  8. Consistency of the maximum likelihood estimator for general hidden Markov models. the Annals of Statistics, 39(1):474–513.
  9. Asymptotic properties of the maximum likelihood estimator in autoregressive models with Markov regime. The Annals of Statistics, 32(5):2254 – 2304.
  10. Nonlinear time series: theory, methods and applications with R examples. Chapman and Hall/CRC.
  11. Identifiability and consistent estimation of nonparametric translation hidden Markov models with general state space. Journal of Machine Learning Research, 21(115):1–40.
  12. Nonparametric finite translation hidden Markov models and extensions. Bernoulli, 22(1):193 – 212.
  13. Modeling and forecasting electricity prices with input/output hidden markov models. IEEE Transactions on Power Systems, 20(1):13–24.
  14. Long short-term memory. Neural computation, 9(8):1735–1780.
  15. Holt, C. C. (2004). Forecasting seasonals and trends by exponentially weighted moving averages. International Journal of Forecasting, 20:5–10.
  16. Forecasting: Principles and Practice. OTexts, Australia, 3nd edition.
  17. Package ‘forecast’. Online] https://cran. r-project. org/web/packages/forecast/forecast. pdf.
  18. Mixture autoregressive hidden Markov models for speech signals. IEEE Transactions on Acoustics, Speech, and Signal Processing, 33(6):1404–1413.
  19. Modeling long- and short-term temporal patterns with deep neural networks.
  20. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. arXiv preprint arXiv:1907.00235.
  21. Temporal Fusion Transformers for interpretable multi-horizon time series forecasting. International Journal of Forecasting, 37(4):1748–1764.
  22. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In International Conference on Learning Representations.
  23. Forecasting time series with complex seasonal patterns using exponential smoothing. Journal of the American Statistical Association, 106(496):1513–1527.
  24. The M4 competition: Results, findings, conclusion and way forward. International Journal of Forecasting, 34(4):802–808.
  25. M5 accuracy competition: Results, findings, and conclusions. International Journal of Forecasting, 38(4):1346–1364.
  26. A time series is worth 64 words: Long-term forecasting with transformers.
  27. NeuralForecast: User friendly state-of-the-art neural forecasting models. PyCon Salt Lake City, Utah, US 2022.
  28. N-beats: Neural basis expansion analysis for interpretable time series forecasting. arXiv preprint arXiv:1905.10437.
  29. Contextual hidden Markov models. In 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2113–2116. IEEE.
  30. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. International Journal of Forecasting, 36(3):1181–1191.
  31. Särkkä, S. (2013). Bayesian filtering and smoothing. Cambridge university press.
  32. Forecasting at scale. The American Statistician, 72.
  33. Touron, A. (2019). Consistency of the maximum likelihood estimator in seasonal hidden Markov models. Statistics and Computing, 29(5):1055–1075.
  34. Trindade, A. (2015). ElectricityLoadDiagrams20112014. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C58C86.
  35. Neural discrete representation learning.
  36. Attention is all you need. 31st Conference on Neural Information Processing Systems (NeurIPS 2017).
  37. Etsformer: Exponential smoothing transformers for time-series forecasting.
  38. Learning deep time-index models for time series forecasting.
  39. Timesnet: Temporal 2d-variation modeling for general time series analysis.
  40. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting.
  41. Are transformers effective for time series forecasting?
  42. Informer: Beyond efficient transformer for long sequence time-series forecasting.
  43. Informer: Beyond efficient transformer for long sequence time-series forecasting. In The Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Virtual Conference, volume 35, pages 11106–11115. AAAI Press.
  44. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting.
Citations (1)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube