Time Series Forecasting with LLMs: Understanding and Enhancing Model Capabilities (2402.10835v4)
Abstract: LLMs have been applied in many fields and have developed rapidly in recent years. As a classic machine learning task, time series forecasting has recently been boosted by LLMs. Recent works treat LLMs as \emph{zero-shot} time series reasoners without further fine-tuning, which achieves remarkable performance. However, there are some unexplored research problems when applying LLMs for time series forecasting under the zero-shot setting. For instance, the LLMs' preferences for the input time series are less understood. In this paper, by comparing LLMs with traditional time series forecasting models, we observe many interesting properties of LLMs in the context of time series forecasting. First, our study shows that LLMs perform well in predicting time series with clear patterns and trends, but face challenges with datasets lacking periodicity. This observation can be explained by the ability of LLMs to recognize the underlying period within datasets, which is supported by our experiments. In addition, the input strategy is investigated, and it is found that incorporating external knowledge and adopting natural language paraphrases substantially improve the predictive performance of LLMs for time series. Overall, our study contributes insight into LLMs' advantages and limitations in time series forecasting under different conditions.
- George EP Box and David A Pierce. 1970. Distribution of residual autocorrelations in autoregressive-integrated moving average time series models. Journal of the American statistical Association, 65(332):1509–1526.
- Mohammad Braei and Sebastian Wagner. 2020. Anomaly detection in univariate time-series: A survey on the state-of-the-art. arXiv preprint arXiv:2004.00433.
- Tempo: Prompt-based generative pre-trained transformer for time series forecasting. arXiv preprint arXiv:2310.04948.
- Llm4ts: Aligning pre-trained llms as data-efficient time-series forecasters.
- Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555.
- Stl: A seasonal-trend decomposition. J. Off. Stat, 6(1):3–73.
- A decoder-only foundation model for time-series forecasting. arXiv preprint arXiv:2310.10688.
- Everette S Gardner Jr. 2006. Exponential smoothing: The state of the art—part ii. International journal of forecasting, 22(4):637–666.
- Azul Garza and Max Mergenthaler-Canseco. 2023. Timegpt-1. arXiv preprint arXiv:2310.03589.
- Bidirectional lstm networks for improved phoneme classification and recognition. In International conference on artificial neural networks, pages 799–804. Springer.
- Large language models are zero-shot time series forecasters. arXiv preprint arXiv:2310.07820.
- Finbert: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2):806–841.
- Hugging Face. 2023. Chapter 6.5 of nlp course. https://huggingface.co/learn/nlp-course/chapter6/5. Accessed: 2023-02-10.
- Time-llm: Time series forecasting by reprogramming large language models. arXiv preprint arXiv:2310.01728.
- Artificial intelligence in customer relationship management: literature review and future research directions. Journal of Business & Industrial Marketing, 37(13):48–63.
- Long short term memory networks for anomaly detection in time series. In Esann, volume 2015, page 89.
- Deepant: A deep learning approach for unsupervised anomaly detection in time series. Ieee Access, 7:1991–2005.
- How to construct deep recurrent neural networks. arXiv preprint arXiv:1312.6026.
- A study of generative large language model for medical research and healthcare. arXiv preprint arXiv:2305.13523.
- Lag-llama: Towards foundation models for time series forecasting. arXiv preprint arXiv:2310.08278.
- Numeric magnitude comparison effects in large language models. In Findings of the Association for Computational Linguistics: ACL 2023, pages 6147–6161, Toronto, Canada. Association for Computational Linguistics.
- Timeseries anomaly detection using temporal hierarchical one-class network. Advances in Neural Information Processing Systems, 33:13016–13026.
- Harmanjeet Singh and Manisha Malhotra. 2023. A time series analysis-based stock price prediction framework using artificial intelligence. In International Conference on Artificial Intelligence of Things, pages 280–289. Springer.
- Test: Text prototype aligned embedding to activate llm’s ability for time series. arXiv preprint arXiv:2308.08241.
- Characteristic-based clustering for time series data. Data mining and knowledge Discovery, 13:335–364.
- Hao Xue and Flora D Salim. 2023. Promptcast: A new prompt-based learning paradigm for time series forecasting. IEEE Transactions on Knowledge and Data Engineering.
- Mingyu Jin (38 papers)
- Hua Tang (6 papers)
- Chong Zhang (137 papers)
- Qinkai Yu (10 papers)
- Yongfeng Zhang (163 papers)
- Mengnan Du (90 papers)
- Zhenting Wang (41 papers)
- Xiaobo Jin (26 papers)