Blending gradient boosted trees and neural networks for point and probabilistic forecasting of hierarchical time series (2310.13029v1)
Abstract: In this paper we tackle the problem of point and probabilistic forecasting by describing a blending methodology of machine learning models that belong to gradient boosted trees and neural networks families. These principles were successfully applied in the recent M5 Competition on both Accuracy and Uncertainty tracks. The keypoints of our methodology are: a) transform the task to regression on sales for a single day b) information rich feature engineering c) create a diverse set of state-of-the-art machine learning models and d) carefully construct validation sets for model tuning. We argue that the diversity of the machine learning models along with the careful selection of validation examples, where the most important ingredients for the effectiveness of our approach. Although forecasting data had an inherent hierarchy structure (12 levels), none of our proposed solutions exploited that hierarchical scheme. Using the proposed methodology, our team was ranked within the gold medal range in both Accuracy and the Uncertainty track. Inference code along with already trained models are available at https://github.com/IoannisNasios/M5_Uncertainty_3rd_place
- The m5 accuracy competition: Results, findings and conclusions. 2020, URL: https://www. researchgate. net/publication/344487258_The_M5_ Accuracy_competition_Results_findings_and_conclusions (2022).
- The m4 competition: 100,000 time series and 61 forecasting methods, International Journal of Forecasting 36 (2020) 54–74. URL: https://www.sciencedirect.com/science/article/pii/S0169207019301128. doi:https://doi.org/10.1016/j.ijforecast.2019.04.014, m4 Competition.
- Statistical and machine learning forecasting methods: Concerns and ways forward, PloS one 13 (2018) e0194889.
- Box office forecasting using machine learning algorithms based on sns data, International Journal of Forecasting 31 (2015) 364–390.
- Advances in forecasting with neural networks? empirical evidence from the nn3 competition on time series prediction, International Journal of Forecasting 27 (2011) 635–660.
- J. H. Friedman, Greedy function approximation: a gradient boosting machine, Annals of statistics (2001) 1189–1232.
- J. H. Friedman, Stochastic gradient boosting, Computational Statistics & Data Analysis 38 (2002) 367–378.
- A comparative study of categorical variable encoding techniques for neural network classifiers, International journal of computer applications 175 (2017) 7–9.
- Polyglot: Distributed word representations for multilingual nlp, arXiv preprint arXiv:1307.1662 (2013).
- Flair: An easy-to-use framework for state-of-the-art nlp, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations), 2019, pp. 54–59.
- Y. Bengio, Y. Grandvalet, No unbiased estimator of the variance of k-fold cross-validation, Journal of machine learning research 5 (2004) 1089–1105.
- R. J. Hyndman, A. B. Koehler, Another look at measures of forecast accuracy, International Journal of Forecasting 22 (2006) 679–688.
- Lightgbm: A highly efficient gradient boosting decision tree, in: Advances in Neural Information Processing Systems, 2017, pp. 3146–3154.
- Stochastic gradient boosted distributed decision trees, in: Proceedings of the 18th ACM Conference on Information and Knowledge Management, 2009, pp. 2061–2064.
- M. C. Tweedie, et al., Statistical properties of inverse gaussian distributions. i, Annals of Mathematical Statistics 28 (1957) 362–377.
- Tweedie gradient boosting for extremely unbalanced zero-inflated data, Communications in Statistics-Simulation and Computation (2020) 1–23.
- Tensorflow: A system for large-scale machine learning, in: 12th {{\{{USENIX}}\}} symposium on operating systems design and implementation ({{\{{OSDI}}\}} 16), 2016, pp. 265–283.
- J. Howard, S. Gugger, Fastai: A layered api for deep learning, Information 11 (2020) 108.
- Pytorch: An imperative style, high-performance deep learning library, arXiv preprint arXiv:1912.01703 (2019).
- Optimally reconciling forecasts in a hierarchy, Foresight: The International Journal of Applied Forecasting (2014) 42–48.