Understanding the Role of Textual Prompts in LLM for Time Series Forecasting: an Adapter View (2311.14782v2)
Abstract: In the burgeoning domain of LLMs, there is a growing interest in applying LLM to time series forecasting, with multiple studies focused on leveraging textual prompts to further enhance the predictive prowess. This study aims to understand how and why the integration of textual prompts into LLM can effectively improve the prediction accuracy of time series, which is not obvious at the glance, given the significant domain gap between texts and time series. Our extensive examination leads us to believe that (a) adding text prompts is roughly equivalent to introducing additional adapters, and (b) It is the introduction of learnable parameters rather than textual information that aligns the LLM with the time series forecasting task, ultimately enhancing prediction accuracy. Inspired by this discovery, we developed four adapters that explicitly address the gap between LLM and time series, and further improve the prediction accuracy. Overall,our work highlights how textual prompts enhance LLM accuracy in time series forecasting and suggests new avenues for continually improving LLM-based time series analysis.
- Tian Zhou (57 papers)
- Xue Wang (69 papers)
- Liang Sun (124 papers)
- Rong Jin (164 papers)
- PeiSong Niu (4 papers)