Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Online Variational Sequential Monte Carlo (2312.12616v3)

Published 19 Dec 2023 in stat.ML and cs.LG

Abstract: Being the most classical generative model for serial data, state-space models (SSM) are fundamental in AI and statistical machine learning. In SSM, any form of parameter learning or latent state inference typically involves the computation of complex latent-state posteriors. In this work, we build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference by combining particle methods and variational inference. While standard VSMC operates in the offline mode, by re-processing repeatedly a given batch of data, we distribute the approximation of the gradient of the VSMC surrogate ELBO in time using stochastic approximation, allowing for online learning in the presence of streams of data. This results in an algorithm, online VSMC, that is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation. In addition, we provide rigorous theoretical results describing the algorithm's convergence properties as the number of data tends to infinity as well as numerical illustrations of its excellent convergence properties and usefulness also in batch-processing settings.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Bishop, C. M. (2016). Pattern Recognition and Machine Learning. Springer.
  2. Importance weighted autoencoders. In International Conference on Learning Representations.
  3. Cappé, O. (2011). Online EM algorithm for hidden Markov models. J. Comput. Graph. Statist., 20(3):728–749.
  4. Inference in Hidden Markov Models. Springer, New York.
  5. An introduction to sequential Monte Carlo methods. Springer, New York.
  6. A recurrent latent variable model for sequential data. In Advances in Neural Information Processing Systems, volume 28, pages 2980 – 2988. Curran Associates, Inc.
  7. Adaptive methods for sequential importance sampling with application to state space models. Stat. Comput., 18(4):461–480.
  8. Adaptive sequential Monte Carlo by means of mixture of experts. Stat. Comput., 24(3):317–337.
  9. Del Moral, P. (2004). Feynman–Kac Formulae. Genealogical and Interacting Particle Systems with Applications. Springer, New York.
  10. Uniform stability of a particle approximation of the optimal filter derivative. SIAM Journal on Control and Optimization, 53(3):1278–1304.
  11. Markov Chains. Spinger.
  12. Sequential Monte Carlo Methods in Practice. Springer, New York.
  13. On sequential Monte-Carlo sampling methods for Bayesian filtering. Stat. Comput., 10:197–208.
  14. Novel approach to nonlinear/non-Gaussian Bayesian state estimation. IEE Proc. F, Radar Signal Process., 140:107–113.
  15. Neural adaptive sequential Monte Carlo. Advances in neural information processing systems, 28.
  16. The pricing of options on assets with stochastic volatilities. J. Finance, 42:281–300.
  17. On particle methods for parameter estimation in state-space models. Statistical science, 30(3):328–351.
  18. Non-asymptotic analysis of biased stochastic approximation scheme. In Conference on Learning Theory, pages 1944–1974. PMLR.
  19. Adam: A method for stochastic optimization. In International Conference on Learning Representations.
  20. Auto-encoding variational Bayes. In International Conference on Learning Representations.
  21. Kitagawa, G. (1996). Monte-Carlo filter and smoother for non-Gaussian nonlinear state space models. J. Comput. Graph. Statist., 1:1–25.
  22. Auto-encoding sequential Monte Carlo. In International Conference on Learning Representations.
  23. Recursive estimation in HMMs. In Proc. IEEE Conf. Decis. Control, pages 3468–3473.
  24. Filtering variational objectives. Advances in Neural Information Processing Systems, 30.
  25. Markov Chains and Stochastic Stability. Cambridge University Press, London.
  26. Online learning with hidden Markov models. Neural Computation, 20(7):1706–1716.
  27. Variational sequential Monte Carlo. In International conference on artificial intelligence and statistics, pages 968–977. PMLR.
  28. Particle-based online estimation of tangent filters with application to parameter estimation in nonlinear state-space models. Annals of the Institute of Statistical Mathematics, 72:545–576.
  29. Particle approximations of the score and observed information matrix in state space models with application to parameter estimation. Biometrika, 98(1):65–80.
  30. A stochastic approximation method. Ann. Math. Statist., 22:400–407.
  31. Asymptotic properties of recursive maximum likelihood estimation in non-linear state-space models. arXiv preprint arXiv:1806.09571.
  32. Streaming variational monte carlo. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):1150–1161.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets