Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 84 tok/s
Gemini 2.5 Pro 45 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 21 tok/s Pro
GPT-4o 92 tok/s Pro
GPT OSS 120B 425 tok/s Pro
Kimi K2 157 tok/s Pro
2000 character limit reached

NFCL: Simply interpretable neural networks for a short-term multivariate forecasting (2405.13393v1)

Published 22 May 2024 in cs.LG and cs.AI

Abstract: Multivariate time-series forecasting (MTSF) stands as a compelling field within the machine learning community. Diverse neural network based methodologies deployed in MTSF applications have demonstrated commendable efficacy. Despite the advancements in model performance, comprehending the rationale behind the model's behavior remains an enigma. Our proposed model, the Neural ForeCasting Layer (NFCL), employs a straightforward amalgamation of neural networks. This uncomplicated integration ensures that each neural network contributes inputs and predictions independently, devoid of interference from other inputs. Consequently, our model facilitates a transparent explication of forecast results. This paper introduces NFCL along with its diverse extensions. Empirical findings underscore NFCL's superior performance compared to nine benchmark models across 15 available open datasets. Notably, NFCL not only surpasses competitors but also provides elucidation for its predictions. In addition, Rigorous experimentation involving diverse model structures bolsters the justification of NFCL's unique configuration.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Neural additive models: Interpretable machine learning with neural nets. Advances in Neural Information Processing Systems 34 (2021).
  2. Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. Journal of Big Data 8, 1 (31 Mar 2021), 53. https://doi.org/10.1186/s40537-021-00444-8
  3. TimeSHAP: Explaining Recurrent Models through Sequence Perturbations. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (Virtual Event, Singapore) (KDD ’21). Association for Computing Machinery, New York, NY, USA, 2565–2573. https://doi.org/10.1145/3447548.3467166
  4. A new accuracy measure based on bounded relative error for time series forecasting. PLoS One 12, 3 (March 2017), e0174202.
  5. Song Chen. 2019. Beijing Multi-Site Air-Quality Data. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5RK5G.
  6. Long sequence time-series forecasting with deep learning: A survey. Information Fusion 97 (2023), 101819. https://doi.org/10.1016/j.inffus.2023.101819
  7. François Chollet. 2016. Xception: Deep Learning with Depthwise Separable Convolutions. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 1800–1807.
  8. James W. Cooley and John W. Tukey. 1965. An Algorithm for the Machine Calculation of Complex Fourier Series. Math. Comp. 19, 90 (1965), 297–301.
  9. Long-term Forecasting with TiDE: Time-series Dense Encoder. Transactions on Machine Learning Research (2023).
  10. Scaling Vision Transformers to 22 Billion Parameters. In Proceedings of the 40th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 202), Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (Eds.). PMLR, 7480–7512.
  11. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations.
  12. S.L. Ho and M. Xie. 1998. The use of ARIMA models for reliability forecasting and analysis. Computers & Industrial Engineering 35, 1 (1998), 213–216. https://doi.org/10.1016/S0360-8352(98)00066-7
  13. Short-term forecasting of passenger demand under on-demand ride services: A spatio-temporal deep learning approach. Transportation Research Part C: Emerging Technologies 85 (2017), 591–608. https://doi.org/10.1016/j.trc.2017.10.016
  14. Short-term prediction of particulate matter (PM10 and PM2.5) in Seoul, South Korea using tree-based machine learning algorithms. Atmospheric Pollution Research 13, 10 (2022), 101547. https://doi.org/10.1016/j.apr.2022.101547
  15. Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (Ann Arbor, MI, USA) (SIGIR ’18). Association for Computing Machinery, New York, NY, USA, 95–104. https://doi.org/10.1145/3209978.3210006
  16. Temporal Convolutional Networks: A Unified Approach to Action Segmentation. In Computer Vision – ECCV 2016 Workshops, Gang Hua and Hervé Jégou (Eds.). Springer International Publishing, Cham, 47–54.
  17. Ke Li and Kaixu Bai. 2019. Spatiotemporal Associations between PM(2.5) and SO(2) as well as NO(2) in China from 2015 to 2018. Int J Environ Res Public Health 16, 13 (July 2019).
  18. SCINet: Time Series Modeling and Forecasting with Sample Convolution and Interaction. In Advances in Neural Information Processing Systems, S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh (Eds.), Vol. 35. Curran Associates, Inc., 5816–5828.
  19. Short-term Load Forecasting Based on GBDT Combinatorial Optimization. In 2018 2nd IEEE Conference on Energy Internet and Energy System Integration (EI2). 1–5. https://doi.org/10.1109/EI2.2018.8582108
  20. Pyraformer: Low-Complexity Pyramidal Attention for Long-Range Time Series Modeling and Forecasting. In International Conference on Learning Representations.
  21. iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. In The Twelfth International Conference on Learning Representations.
  22. Ilya Loshchilov and Frank Hutter. 2019. Decoupled Weight Decay Regularization. In International Conference on Learning Representations.
  23. Scott M Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.). Curran Associates, Inc., 4765–4774.
  24. Short-Term Load Forecasting on Individual Consumers. Energies 15, 16 (2022). https://doi.org/10.3390/en15165856
  25. Short-term electric load forecasting using an EMD-BI-LSTM approach for smart grid energy management system. Energy and Buildings 288 (2023), 113022. https://doi.org/10.1016/j.enbuild.2023.113022
  26. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In International Conference on Learning Representations.
  27. PyTorch: an imperative style, high-performance deep learning library. Curran Associates Inc., Red Hook, NY, USA.
  28. W. C. Porter and C. L. Heald. 2019. The mechanisms and meteorological drivers of the summertime ozone–temperature relationship. Atmospheric Chemistry and Physics 19, 21 (2019), 13367–13381. https://doi.org/10.5194/acp-19-13367-2019
  29. Neural Basis Models for Interpretability. arXiv:2205.14120 (2022).
  30. Explainable Artificial Intelligence (XAI) on TimeSeries Data: A Survey. CoRR abs/2104.00950 (2021). arXiv:2104.00950
  31. Iqbal H. Sarker. 2021. Deep Learning: A Comprehensive Overview on Techniques, Taxonomy, Applications and Research Directions. SN Computer Science 2, 6 (18 Aug 2021), 420. https://doi.org/10.1007/s42979-021-00815-1
  32. Li Sun and Fengqi You. 2021. Machine Learning and Data-Driven Techniques for the Control of Smart Power Generation Systems: An Uncertainty Handling Perspective. Engineering 7, 9 (2021), 1239–1247. https://doi.org/10.1016/j.eng.2021.04.020
  33. Mingxing Tan and Quoc Le. 2021. EfficientNetV2: Smaller Models and Faster Training. In Proceedings of the 38th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 139), Marina Meila and Tong Zhang (Eds.). PMLR, 10096–10106.
  34. Explainable AI for Time Series Classification: A Review, Taxonomy and Research Directions. IEEE Access 10 (2022), 100700–100724. https://doi.org/10.1109/ACCESS.2022.3207765
  35. J V Tu. 1996. Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes. J Clin Epidemiol 49, 11 (Nov. 1996), 1225–1231.
  36. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (Long Beach, California, USA) (NIPS’17). Curran Associates Inc., Red Hook, NY, USA, 6000–6010.
  37. Validation of XAI explanations for multivariate time series classification in the maritime domain. Journal of Computational Science 58 (2022), 101539. https://doi.org/10.1016/j.jocs.2021.101539
  38. DWFH: An improved data-driven deep weather forecasting hybrid model using Transductive Long Short Term Memory (T-LSTM). Expert Systems with Applications 213 (2023), 119270. https://doi.org/10.1016/j.eswa.2022.119270
  39. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. In International Conference on Learning Representations.
  40. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. In Advances in Neural Information Processing Systems.
  41. Understanding and Improving Layer Normalization. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc.
  42. Hiroshi Yoshikado. 2023. Correlation between air temperature and surface ozone in their extreme ranges in the greater Tokyo region. Asian Journal of Atmospheric Environment 17, 1 (Aug. 2023), 9.
  43. Are Transformers Effective for Time Series Forecasting? Proceedings of the AAAI Conference on Artificial Intelligence.
  44. A novel DWTimesNet-based short-term multi-step wind power forecasting model using feature selection and auto-tuning methods. Energy Conversion and Management 301 (2024), 118045. https://doi.org/10.1016/j.enconman.2023.118045
  45. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In AAAI Conference on Artificial Intelligence.
  46. FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting. In Proc. 39th International Conference on Machine Learning (ICML 2022) (Baltimore, Maryland).
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Authors (2)

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets