Add and Thin: Diffusion for Temporal Point Processes (2311.01139v2)
Abstract: Autoregressive neural networks within the temporal point process (TPP) framework have become the standard for modeling continuous-time event data. Even though these models can expressively capture event sequences in a one-step-ahead fashion, they are inherently limited for long-term forecasting applications due to the accumulation of errors caused by their sequential nature. To overcome these limitations, we derive ADD-THIN, a principled probabilistic denoising diffusion model for TPPs that operates on entire event sequences. Unlike existing diffusion approaches, ADD-THIN naturally handles data with discrete and continuous components. In experiments on synthetic and real-world datasets, our model matches the state-of-the-art TPP models in density estimation and strongly outperforms them in forecasting.
- J. M. L. Alcaraz and N. Strodthoff. Diffusion-based time series imputation and forecasting with structured state space models. arXiv preprint arXiv:2208.09399, 2022.
- Structured denoising diffusion models in discrete state-spaces. Advances in Neural Information Processing Systems, 34:17981–17993, 2021.
- Modeling temporal data as continuous functions with stochastic process diffusion. In International Conference on Machine Learning (ICML), 2023.
- T. Bosser and S. B. Taieb. On the predictive accuracy of neural temporal point process models for continuous-time event data. Transactions on Machine Learning Research, 2023.
- Neural spatio-temporal point processes. arXiv preprint arXiv:2011.04583, 2020.
- T. Chen and M. Zhou. Learning to jump: Thinning and thickening latent counts for generative modeling. arXiv preprint arXiv:2305.18375, 2023.
- D. R. Cox. Some statistical methods connected with series of events. Journal of the Royal Statistical Society: Series B (Methodological), 17(2):129–157, 1955.
- D. Daley and D. Vere-Jones. An Introduction to the Theory of Point Processes: Volume I: Elementary Theory and Methods. Probability and Its Applications. Springer New York, 2006.
- D. J. Daley and D. Vere-Jones. An introduction to the theory of point processes: volume II: general theory and structure. Springer Science & Business Media, 2007.
- Using deep learning for flexible and scalable earthquake forecasting. Geophysical Research Letters, 50(17), 2023.
- C. DR and I. COLL. The statistical analysis of dependencies in point processes. Stochastic Point Processes. Wiley: New York, pages 55–66, 1972.
- Recurrent marked temporal point processes: Embedding event history to vector. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1555–1564, 2016.
- Neural temporal point processes for modelling electronic health records. In Machine Learning for Health, pages 85–113. PMLR, 2020.
- A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723–773, 2012.
- Card: Classification and regression diffusion models. arXiv preprint arXiv:2206.07275, 2022.
- Inference and sampling of point processes from diffusion excursions. In The 39th Conference on Uncertainty in Artificial Intelligence, 2023.
- A. G. Hawkes. Spectra of some self-exciting and mutually exciting point processes. Biometrika, 58(1):83–90, 1971.
- Denoising diffusion probabilistic models. Neural Information Processing Systems (NeurIPS), 2020.
- C. Hong and C. Shelton. Deep neyman-scott processes. In Proceedings of the 25th International Conference on Artificial Intelligence and Statistics, volume 151, pages 3627–3646. PMLR, 2022.
- V. Isham and M. Westcott. A self-correcting point process. Stochastic processes and their applications, 8(3):335–347, 1979.
- J. Jia and A. R. Benson. Neural jump stochastic differential equations. Advances in Neural Information Processing Systems, 32, 2019.
- D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Turbulent flow simulation using autoregressive conditional diffusion models, 2023.
- Predict, refine, synthesize: Self-guiding diffusion models for probabilistic time series forecasting. arXiv preprint arXiv:2307.11494, 2023.
- Srdiff: Single image super-resolution with diffusion probabilistic models. Neurocomputing, 479:47–59, 2022a.
- Learning temporal point processes via reinforcement learning. Advances in neural information processing systems, 31, 2018.
- Diffusion-lm improves controllable text generation. Advances in Neural Information Processing Systems, 35:4328–4343, 2022b.
- From zero to turbulence: Generative modeling for 3d flow simulation, 2023.
- Exploring generative neural temporal point process. Transactions on Machine Learning Research, 2022.
- Pde-refiner: Achieving accurate long rollouts with neural pde solvers. arXiv preprint arXiv:2308.05732, 2023.
- S. Luo and W. Hu. Diffusion probabilistic models for 3d point cloud generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2837–2845, 2021.
- A conditional point diffusion-refinement paradigm for 3d point cloud completion. arXiv preprint arXiv:2112.03530, 2021.
- H. Mei and J. M. Eisner. The neural hawkes process: A neurally self-modulating multivariate point process. In Neural Information Processing Systems (NeurIPS), 2017.
- Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741, 2021.
- A. Q. Nichol and P. Dhariwal. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pages 8162–8171. PMLR, 2021.
- Deep mixture point processes: Spatio-temporal event prediction with rich contextual information. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 373–383, 2019.
- Fully neural network based model for general temporal point processes. Advances in neural information processing systems, 32, 2019.
- On wasserstein two-sample testing and related families of nonparametric tests. Entropy, 19(2):47, 2017.
- B. Sevast’yanov. Renewal theory. Journal of Soviet Mathematics, 4(3):281–302, 1975.
- Identifying coordinated accounts on social media through hidden influence and group behaviours. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pages 1441–1451, 2021.
- Intensity-free learning of temporal point processes. In International Conference on Learning Representations (ICLR), 2020a.
- Fast and flexible temporal point processes with triangular maps. In Advances in Neural Information Processing Systems (NeurIPS), 2020b.
- Neural temporal point processes: A review. arXiv preprint arXiv:2104.03528, 2021.
- Unipoint: Universally approximating point processes intensities. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 9685–9694, 2021.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning (ICML), pages 2256–2265, 2015.
- Csdi: Conditional score-based diffusion models for probabilistic time series imputation. Advances in Neural Information Processing Systems, 34:24804–24816, 2021.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Digress: Discrete denoising diffusion for graph generation. arXiv preprint arXiv:2209.14734, 2022.
- Wasserstein learning of deep generative point process models. Advances in neural information processing systems, 30, 2017.
- Self-attentive hawkes process. In International conference on machine learning, pages 11183–11193. PMLR, 2020a.
- Cause: Learning granger causality from event sequences using attribution methods. In International Conference on Machine Learning, pages 11235–11245. PMLR, 2020b.
- Transformer hawkes process. arXiv preprint arXiv:2002.09291, 2020.