Variational quantization for state space models (2404.11117v1)
Abstract: Forecasting tasks using large datasets gathering thousands of heterogeneous time series is a crucial statistical problem in numerous sectors. The main challenge is to model a rich variety of time series, leverage any available external signals and provide sharp predictions with statistical guarantees. In this work, we propose a new forecasting model that combines discrete state space hidden Markov models with recent neural network architectures and training procedures inspired by vector quantized variational autoencoders. We introduce a variational discrete posterior distribution of the latent states given the observations and a two-stage training procedure to alternatively train the parameters of the latent states and of the emission distributions. By learning a collection of emission laws and temporarily activating them depending on the hidden process dynamics, the proposed method allows to explore large datasets and leverage available external signals. We assess the performance of the proposed method using several datasets and show that it outperforms other state-of-the-art solutions.
- An Input Output HMM architecture. In Tesauro, G., Touretzky, D., and Leen, T., editors, Advances in Neural Information Processing Systems, volume 7. MIT Press.
- Time series analysis: forecasting and control. John Wiley & Sons.
- The fundamental theorem of exponential smoothing. Operations Research, 9(5):673–685.
- Nhits: Neural hierarchical interpolation for time series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 6989–6997.
- An introduction to sequential Monte Carlo. Springer.
- HERMES: Hybrid error-corrector model with inclusion of external signals for nonstationary fashion time series. preprint.
- Identifiability of discrete Input-output hidden Markov models with external signals. working paper or preprint.
- Consistency of the maximum likelihood estimator for general hidden Markov models. the Annals of Statistics, 39(1):474–513.
- Asymptotic properties of the maximum likelihood estimator in autoregressive models with Markov regime. The Annals of Statistics, 32(5):2254 – 2304.
- Nonlinear time series: theory, methods and applications with R examples. Chapman and Hall/CRC.
- Identifiability and consistent estimation of nonparametric translation hidden Markov models with general state space. Journal of Machine Learning Research, 21(115):1–40.
- Nonparametric finite translation hidden Markov models and extensions. Bernoulli, 22(1):193 – 212.
- Modeling and forecasting electricity prices with input/output hidden markov models. IEEE Transactions on Power Systems, 20(1):13–24.
- Long short-term memory. Neural computation, 9(8):1735–1780.
- Holt, C. C. (2004). Forecasting seasonals and trends by exponentially weighted moving averages. International Journal of Forecasting, 20:5–10.
- Forecasting: Principles and Practice. OTexts, Australia, 3nd edition.
- Package ‘forecast’. Online] https://cran. r-project. org/web/packages/forecast/forecast. pdf.
- Mixture autoregressive hidden Markov models for speech signals. IEEE Transactions on Acoustics, Speech, and Signal Processing, 33(6):1404–1413.
- Modeling long- and short-term temporal patterns with deep neural networks.
- Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. arXiv preprint arXiv:1907.00235.
- Temporal Fusion Transformers for interpretable multi-horizon time series forecasting. International Journal of Forecasting, 37(4):1748–1764.
- Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In International Conference on Learning Representations.
- Forecasting time series with complex seasonal patterns using exponential smoothing. Journal of the American Statistical Association, 106(496):1513–1527.
- The M4 competition: Results, findings, conclusion and way forward. International Journal of Forecasting, 34(4):802–808.
- M5 accuracy competition: Results, findings, and conclusions. International Journal of Forecasting, 38(4):1346–1364.
- A time series is worth 64 words: Long-term forecasting with transformers.
- NeuralForecast: User friendly state-of-the-art neural forecasting models. PyCon Salt Lake City, Utah, US 2022.
- N-beats: Neural basis expansion analysis for interpretable time series forecasting. arXiv preprint arXiv:1905.10437.
- Contextual hidden Markov models. In 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2113–2116. IEEE.
- DeepAR: Probabilistic forecasting with autoregressive recurrent networks. International Journal of Forecasting, 36(3):1181–1191.
- Särkkä, S. (2013). Bayesian filtering and smoothing. Cambridge university press.
- Forecasting at scale. The American Statistician, 72.
- Touron, A. (2019). Consistency of the maximum likelihood estimator in seasonal hidden Markov models. Statistics and Computing, 29(5):1055–1075.
- Trindade, A. (2015). ElectricityLoadDiagrams20112014. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C58C86.
- Neural discrete representation learning.
- Attention is all you need. 31st Conference on Neural Information Processing Systems (NeurIPS 2017).
- Etsformer: Exponential smoothing transformers for time-series forecasting.
- Learning deep time-index models for time series forecasting.
- Timesnet: Temporal 2d-variation modeling for general time series analysis.
- Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting.
- Are transformers effective for time series forecasting?
- Informer: Beyond efficient transformer for long sequence time-series forecasting.
- Informer: Beyond efficient transformer for long sequence time-series forecasting. In The Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Virtual Conference, volume 35, pages 11106–11115. AAAI Press.
- Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.