HiPPO-KAN: Efficient KAN Model for Time Series Analysis
Abstract: In this study, we introduces a parameter-efficient model that outperforms traditional models in time series forecasting, by integrating High-order Polynomial Projection (HiPPO) theory into the Kolmogorov-Arnold network (KAN) framework. This HiPPO-KAN model achieves superior performance on long sequence data without increasing parameter count. Experimental results demonstrate that HiPPO-KAN maintains a constant parameter count while varying window sizes and prediction horizons, in contrast to KAN, whose parameter count increases linearly with window size. Surprisingly, although the HiPPO-KAN model keeps a constant parameter count as increasing window size, it significantly outperforms KAN model at larger window sizes. These results indicate that HiPPO-KAN offers significant parameter efficiency and scalability advantages for time series forecasting. Additionally, we address the lagging problem commonly encountered in time series forecasting models, where predictions fail to promptly capture sudden changes in the data. We achieve this by modifying the loss function to compute the MSE directly on the coefficient vectors in the HiPPO domain. This adjustment effectively resolves the lagging problem, resulting in predictions that closely follow the actual time series data. By incorporating HiPPO theory into KAN, this study showcases an efficient approach for handling long sequences with improved predictive accuracy, offering practical contributions for applications in large-scale time series data.
- Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021.
- Deep Learning. Adaptive Computation and Machine Learning series. The MIT Press, 2016.
- Kan: Kolmogorov-arnold networks. arXiv preprint arXiv:2404.19756, 2024.
- Kan 2.0: Kolmogorov-arnold networks meet science. arXiv preprint arXiv:2408.10205, 2024.
- G. P. Zhang. An investigation of neural networks for linear time-series forecasting. Computers & Operations Research, 28(12):1183–1202, 2001.
- Neural networks for short-term load forecasting: A review and evaluation. IEEE Transactions on Power Systems, 16(1):44–55, 2001.
- A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:1506.00019, 2015.
- Learning to forget: Continual prediction with lstm. Neural Computation, 12(10):2451–2471, 2000.
- Unlocking the power of lstm for long term time series forecasting. arXiv preprint arXiv:2408:10006, 2024.
- Deep state space models for time series forecasting. In Advances in Neural Information Processing Systems, pages 7796–7805, 2018.
- Learning interpretable deep state space model for probabilistic time series forecasting. arXiv preprint arXiv:2102.00397, 2021.
- H. Inzirillo. Deep state space recurrent neural networks for time series forecasting. arXiv preprint arXiv:2407.15236, 2024.
- Learning long-term dependencies with gradient descent is difficult. IEEE transactions on neural networks, 5(2):157–166, 1994.
- Hippo: Recurrent memory with optimal polynomial projections. arXiv preprint arXiv:2008.07669, 2020.
- Efficiently modeling long sequences with structured state spaces. arXiv preprint arXiv:2111.00396, 2022.
- On the parametrization and initialization of diagonal state space models. arXiv preprint arXiv:2206.11893, 2022.
- How to train your hippo: State space models with generalized orthogonal basis projections. arXiv preprint arXiv:2206.12037, 2022.
- Legendre memory units: Continuous-time representation in recurrent neural networks. In Advances in Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada, 2019.
- Kolmogorov-arnold networks (kans) for time series analysis. arXiv preprint arXiv:2405.08790, 2024.
- Kolmogorov-arnold networks for time series: Bridging predictive power and interpretability. arXiv preprint arXiv:2406.02496, 2024.
- Tkan: Temporal kolmogorov-arnold networks. arXiv preprint arXiv:2405.07344, 2024.
- A temporal kolmogorov-arnold transformer for time series forecasting. arXiv preprint arXiv:2406.02486, 2024.
- Lagging problem in financial time series forecasting. Neural Computing and Applications, 35:20819–20839, 2023.
- Deep learning and the information bottleneck principle. arXiv preprint arXiv:1503.02406, 2015.
- On the information bottleneck theory of deep learning*. Journal of Statistical Mechanics: Theory and Experiment, 2019(12):124020, 2019.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.