Generative Pre-Trained Transformer for Symbolic Regression Base In-Context Reinforcement Learning (2404.06330v1)
Abstract: The mathematical formula is the human language to describe nature and is the essence of scientific research. Finding mathematical formulas from observational data is a major demand of scientific research and a major challenge of artificial intelligence. This area is called symbolic regression. Originally symbolic regression was often formulated as a combinatorial optimization problem and solved using GP or reinforcement learning algorithms. These two kinds of algorithms have strong noise robustness ability and good Versatility. However, inference time usually takes a long time, so the search efficiency is relatively low. Later, based on large-scale pre-training data proposed, such methods use a large number of synthetic data points and expression pairs to train a Generative Pre-Trained Transformer(GPT). Then this GPT can only need to perform one forward propagation to obtain the results, the advantage is that the inference speed is very fast. However, its performance is very dependent on the training data and performs poorly on data outside the training set, which leads to poor noise robustness and Versatility of such methods. So, can we combine the advantages of the above two categories of SR algorithms? In this paper, we propose \textbf{FormulaGPT}, which trains a GPT using massive sparse reward learning histories of reinforcement learning-based SR algorithms as training data. After training, the SR algorithm based on reinforcement learning is distilled into a Transformer. When new test data comes, FormulaGPT can directly generate a "reinforcement learning process" and automatically update the learning policy in context. Tested on more than ten datasets including SRBench, formulaGPT achieves the state-of-the-art performance in fitting ability compared with four baselines. In addition, it achieves satisfactory results in noise robustness, versatility, and inference efficiency.
- Multiple regression genetic programming. In Proceedings of the 2014 Annual Conference on Genetic and Evolutionary Computation, pages 879–886, New York, NY, USA, 2014. Association for Computing Machinery. ISBN 9781450326629. doi: 10.1145/2576768.2598291. URL https://doi.org/10.1145/2576768.2598291.
- Deep reinforcement learning: A brief survey. IEEE Signal Processing Magazine, 34(6):26–38, 2017.
- Neural symbolic regression that scales. In International Conference on Machine Learning, pages 936–945. PMLR, 2021.
- Optimizing the latent space of generative networks. arXiv preprint arXiv:1707.05776, 2017.
- Confidence level and feeling of knowing in question answering: The weight of inferential processes. Journal of Experimental Psychology: Learning, Memory, and Cognition, 18(1):142, 1992.
- Rémi Coulom. Efficient selectivity and backup operators in monte-carlo tree search. In International conference on computers and games, pages 72–83. Springer, 2006.
- Rethinking symbolic regression: Morphology and adaptability in the context of evolutionary algorithms. In The Eleventh International Conference on Learning Representations, 2022.
- Statistical genetic programming for symbolic regression. Applied Soft Computing, 60:447–469, 2017.
- Hado Hasselt. Double q-learning. Advances in neural information processing systems, 23, 2010.
- Thomas Junk. Confidence level computation for combining searches with small statistics. Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, 434(2-3):435–443, 1999.
- Reinforcement learning: A survey. Journal of artificial intelligence research, 4:237–285, 1996.
- End-to-end symbolic regression with transformers. Advances in Neural Information Processing Systems, 35:10269–10281, 2022.
- Deep generative symbolic regression with monte-carlo-tree-search. arXiv preprint arXiv:2302.11223, 2023.
- Integration of neural network-based symbolic regression in deep learning for scientific discovery. IEEE transactions on neural networks and learning systems, 32(9):4166–4177, 2020.
- John R Koza. Genetic programming as a means for programming computers by natural selection. Statistics and computing, 4:87–112, 1994.
- William B Langdon. Genetic programming and data structures: genetic programming+ data structures= automatic programming! 1998.
- In-context reinforcement learning with algorithm distillation. arXiv preprint arXiv:2210.14215, 2022.
- Set transformer: A framework for attention-based permutation-invariant neural networks. In International conference on machine learning, pages 3744–3753. PMLR, 2019a.
- Set transformer: A framework for attention-based permutation-invariant neural networks. In Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 3744–3753. PMLR, 09–15 Jun 2019b. URL https://proceedings.mlr.press/v97/lee19d.html.
- Metasymnet: A dynamic symbolic regression network capable of evolving into arbitrary formulations. arXiv preprint arXiv:2311.07326, 2023.
- Discovering mathematical formulas from data via gpt-guided monte carlo tree search. arXiv preprint arXiv:2401.14424, 2024a.
- Mmsr: Symbolic regression is a multimodal task. arXiv preprint arXiv:2402.18603, 2024b.
- Yuxi Li. Deep reinforcement learning: An overview. arXiv preprint arXiv:1701.07274, 2017.
- On the limited memory bfgs method for large scale optimization. Mathematical programming, 45(1-3):503–528, 1989.
- Cross-entropy loss functions: Theoretical analysis and applications. In International Conference on Machine Learning, pages 23803–23828. PMLR, 2023.
- Trent McConaghy. FFX: Fast, Scalable, Deterministic Symbolic Regression Technology. Springer New York, New York, NY, 2011. ISBN 978-1-4614-1770-5. doi: 10.1007/978-1-4614-1770-5_13. URL https://doi.org/10.1007/978-1-4614-1770-5_13.
- Snip: Bridging mathematical symbolic and numeric realms with unified pre-training. arXiv preprint arXiv:2310.02227, 2023.
- Human-level control through deep reinforcement learning. nature, 518(7540):529–533, 2015.
- Symbolic regression via neural-guided genetic programming population seeding. arXiv preprint arXiv:2111.00053, 2021.
- Surrogate-assisted genetic programming with simplified models for automated design of dispatching rules. IEEE Transactions on Cybernetics, 47(9):2951–2965, 2017. doi: 10.1109/TCYB.2016.2562674.
- Daniel J Ozer. Correlation and the coefficient of determination. Psychological bulletin, 97(2):307, 1985.
- Deep symbolic regression: Recovering mathematical expressions from data via risk-seeking policy gradients. arXiv preprint arXiv:1912.04871, 2019.
- Alex Sherstinsky. Fundamentals of recurrent neural network (rnn) and long short-term memory (lstm) network. Physica D: Nonlinear Phenomena, 404:132306, 2020.
- Transformer-based planning for symbolic regression. Advances in Neural Information Processing Systems, 36, 2024.
- Symbolic physics learner: Discovering governing equations via monte carlo tree search. arXiv preprint arXiv:2205.13134, 2022.
- Ai feynman: A physics-inspired method for symbolic regression. Science Advances, 6(16):eaay2631, 2020.
- Symbolicgpt: A generative transformer model for symbolic regression. arXiv preprint arXiv:2106.14131, 2021.
- Symformer: End-to-end symbolic regression using transformer-based architecture. IEEE Access, 2024.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Reinforcement symbolic regression machine.
- A review of recurrent neural networks: Lstm cells and network architectures. Neural computation, 31(7):1235–1270, 2019.
- Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems, 31, 2018.