MambaStock: Selective state space model for stock prediction (2402.18959v1)
Abstract: The stock market plays a pivotal role in economic development, yet its intricate volatility poses challenges for investors. Consequently, research and accurate predictions of stock price movements are crucial for mitigating risks. Traditional time series models fall short in capturing nonlinearity, leading to unsatisfactory stock predictions. This limitation has spurred the widespread adoption of neural networks for stock prediction, owing to their robust nonlinear generalization capabilities. Recently, Mamba, a structured state space sequence model with a selection mechanism and scan module (S6), has emerged as a powerful tool in sequence modeling tasks. Leveraging this framework, this paper proposes a novel Mamba-based model for stock price prediction, named MambaStock. The proposed MambaStock model effectively mines historical stock market data to predict future stock prices without handcrafted features or extensive preprocessing procedures. Empirical studies on several stocks indicate that the MambaStock model outperforms previous methods, delivering highly accurate predictions. This enhanced accuracy can assist investors and institutions in making informed decisions, aiming to maximize returns while minimizing risks. This work underscores the value of Mamba in time-series forecasting. Source code is available at https://github.com/zshicode/MambaStock.
- J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” CoRR, 2018.
- A. Gu and T. Dao, “Mamba: Linear-time sequence modeling with selective state spaces,” arXiv preprint arXiv:2312.00752, 2023.
- C. Jin, J. Gao, Z. Shi, and H. Zhang, “Attcry: Attention-based neural network model for protein crystallization prediction,” Neurocomputing, vol. 463, pp. 265–274, 2021.
- C. Jin, Z. Shi, C. Kang, K. Lin, and H. Zhang, “Tlcrys: Transfer learning based method for protein crystallization prediction,” International Journal of Molecular Sciences, vol. 23, p. 972, 1 2022.
- C. Jin, Z. Shi, W. Li, and Y. Guo, “Bidirectional lstm-crf attention-based model for chinese word segmentation,” arXiv preprint arXiv:2105.09681, 2021.
- C. Jin, Z. Shi, K. Lin, and H. Zhang, “Predicting mirna-disease association based on neural inductive matrix completion with graph autoencoders and self-attention mechanism,” Biomolecules, vol. 12, no. 1, p. 64, 2022.
- C. Jin, Z. Shi, H. Zhang, and Y. Yin, “Predicting lncrna–protein interactions based on graph autoencoders and collaborative training,” in IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, USA, 9-12 December, 2021.
- R. E. Kalman, “A new approach to linear filtering and prediction problems,” Journal of Fluids Engineering, vol. 82, no. 1, pp. 35–44, 1960.
- D. P. Kingma and J. Ba, “Adam: a method for stochastic optimization,” in 3rd International Conference for Learning Representations(ICLR), 2015.
- K. Lin, X. Quan, C. Jin, Z. Shi, and J. Yang, “An interpretable double-scale attention model for enzyme protein class prediction based on transformer encoders and multi-scale convolutions,” Frontiers in Genetics, vol. 13, p. 885627, 2022.
- Z. Shi, “Incorporating transformer and lstm to kalman filter with em algorithm for state estimation,” arXiv preprint arXiv:2105.00250, 2021.
- ——, “Differential equation and probability inspired graph neural networks for latent variable learning,” arXiv preprint arXiv:2202.13800, 2022.
- Z. Shi, Y. Hu, G. Mo, and J. Wu, “Attention-based cnn-lstm and xgboost hybrid model for stock prediction,” arXiv preprint arXiv:2204.02623, 2022.
- Z. Shi and B. Li, “Graph neural networks and attention-based cnn-lstm for protein classification,” arXiv preprint arXiv:2204.09486, 2022.
- Z. Shi, H. Zhang, C. Jin, X. Quan, and Y. Yin, “A representation learning model based on variational inference and graph autoencoder for predicting lncrna-disease associations,” BMC Bioinformatics, vol. 22, p. 136, 2021.
- I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural networks,” in NeurIPS, 2014.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” in NeurIPS, 2017.
- G. Zhang, “Time series forecasting using a hybrid arima and neural network model,” Neurocomputing, vol. 50, pp. 159–175, 2003.
- Zhuangwei Shi (7 papers)