Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Recency-Weighted Temporally-Segmented Ensemble for Time-Series Modeling (2403.02150v1)

Published 4 Mar 2024 in stat.ML and cs.LG

Abstract: Time-series modeling in process industries faces the challenge of dealing with complex, multi-faceted, and evolving data characteristics. Conventional single model approaches often struggle to capture the interplay of diverse dynamics, resulting in suboptimal forecasts. Addressing this, we introduce the Recency-Weighted Temporally-Segmented (ReWTS, pronounced roots') ensemble model, a novel chunk-based approach for multi-step forecasting. The key characteristics of the ReWTS model are twofold: 1) It facilitates specialization of models into different dynamics by segmenting the training data intochunks' of data and training one model per chunk. 2) During inference, an optimization procedure assesses each model on the recent past and selects the active models, such that the appropriate mixture of previously learned dynamics can be recalled to forecast the future. This method not only captures the nuances of each period, but also adapts more effectively to changes over time compared to conventional `global' models trained on all data in one go. We present a comparative analysis, utilizing two years of data from a wastewater treatment plant and a drinking water treatment plant in Norway, demonstrating the ReWTS ensemble's superiority. It consistently outperforms the global model in terms of mean squared forecasting error across various model architectures by 10-70\% on both datasets, notably exhibiting greater resilience to outliers. This approach shows promise in developing automatic, adaptable forecasting models for decision-making and control systems in process industries and other complex systems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. B.Wayne Bequette “Process control: modeling, design, and simulation”, Prentice-Hall international series in the physical and chemical engineering sciences Upper Saddle River, NJ: Prentice Hall PTR, 2010
  2. “The future of PID control” In Control Engineering Practice 9.11, PID Control, 2001, pp. 1163–1175
  3. “Review on model predictive control: an engineering perspective” In The International Journal of Advanced Manufacturing Technology 117.5, 2021, pp. 1327–1349 DOI: 10.1007/s00170-021-07682-3
  4. Muhammad Nauman, Wajiha Shireen and Amir Hussain “Model-Free Predictive Control and Its Applications” In Energies 15.14, 2022, pp. 5131 DOI: 10.3390/en15145131
  5. Michael McCloskey and Neal J. Cohen “Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem” In Psychology of Learning and Motivation 24 Elsevier, 1989, pp. 109–165 DOI: 10.1016/S0079-7421(08)60536-8
  6. “Ensemble learning for data stream analysis: A survey” In Information Fusion 37, 2017, pp. 132–156 DOI: 10.1016/j.inffus.2017.02.004
  7. “Mining concept-drifting data streams using ensemble classifiers” In Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining ACM, 2003, pp. 226–235 DOI: 10.1145/956750.956778
  8. “Reacting to Different Types of Concept Drift: The Accuracy Updated Ensemble Algorithm” In IEEE Transactions on Neural Networks and Learning Systems 25.1, 2014, pp. 81–94 DOI: 10.1109/TNNLS.2013.2251352
  9. “Incremental Learning of Concept Drift in Nonstationary Environments” In IEEE Transactions on Neural Networks 22.10, 2011, pp. 1517–1531 DOI: 10.1109/TNN.2011.2160459
  10. “A streaming ensemble algorithm (SEA) for large-scale classification” In Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining, KDD ’01 New York, NY, USA: Association for Computing Machinery, 2001, pp. 377–382 DOI: 10.1145/502512.502568
  11. “Multi-step forecasting for big data time series based on ensemble learning” In Knowledge-Based Systems 163, 2019, pp. 830–841 DOI: 10.1016/j.knosys.2018.10.009
  12. Jorge Nocedal and Stephen J. Wright “Numerical Optimization” New York, NY: Springer, 2006
  13. M.S. Andersen, J. Dahl and L. Vandenberghe “CVXOPT: A Python package for convex optimization, version 1.1.6” Available at http://cvxopt.org, http://cvxopt.org, 2013
  14. “Mixture of experts: a literature survey” In Artificial Intelligence Review 42 Springer, 2014, pp. 275–293
  15. “Darts: User-Friendly Modern Machine Learning for Time Series” In Journal of Machine Learning Research 23.124, 2022
  16. Omry Yadan “Hydra - A framework for elegantly configuring complex applications”, Github, 2019 URL: https://github.com/facebookresearch/hydra
  17. “Optuna: A next-generation hyperparameter optimization framework” In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, 2019, pp. 2623–2631
  18. Ralf C. Staudemeyer and Eric Rothstein Morris “Understanding LSTM – a tutorial into Long Short-Term Memory Recurrent Neural Networks” arXiv, 2019 DOI: 10.48550/arXiv.1909.09586
  19. “Regularization and Variable Selection Via the Elastic Net” In Journal of the Royal Statistical Society Series B: Statistical Methodology 67.2, 2005, pp. 301–320 DOI: 10.1111/j.1467-9868.2005.00503.x
  20. “XGBoost: A Scalable Tree Boosting System” In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16 New York, NY, USA: Association for Computing Machinery, 2016, pp. 785–794 DOI: 10.1145/2939672.2939785
  21. Shaojie Bai, J.Zico Kolter and Vladlen Koltun “An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling” arXiv, 2018

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets