LLM and Simulation as Bilevel Optimizers: A New Paradigm to Advance Physical Scientific Discovery (2405.09783v1)
Abstract: LLMs have recently gained significant attention in scientific discovery for their extensive knowledge and advanced reasoning capabilities. However, they encounter challenges in effectively simulating observational feedback and grounding it with language to propel advancements in physical scientific discovery. Conversely, human scientists undertake scientific discovery by formulating hypotheses, conducting experiments, and revising theories through observational analysis. Inspired by this, we propose to enhance the knowledge-driven, abstract reasoning abilities of LLMs with the computational strength of simulations. We introduce Scientific Generative Agent (SGA), a bilevel optimization framework: LLMs act as knowledgeable and versatile thinkers, proposing scientific hypotheses and reason about discrete components, such as physics equations or molecule structures; meanwhile, simulations function as experimental platforms, providing observational feedback and optimizing via differentiability for continuous parts, such as physical parameters. We conduct extensive experiments to demonstrate our framework's efficacy in constitutive law discovery and molecular design, unveiling novel solutions that differ from conventional human expectations yet remain coherent upon analysis.
- The impact of large language models on scientific discovery: a preliminary study using gpt-4. arXiv preprint arXiv:2311.07361, 2023.
- Anthropic. Introducing the next generation of claude, 2024. URL https://www.anthropic.com/news/claude-3-family.
- Multiple regression genetic programming. In Proceedings of the 2014 Annual Conference on Genetic and Evolutionary Computation, pp. 879–886, 2014.
- Understanding and simplifying one-shot architecture search. In International conference on machine learning, pp. 550–559. PMLR, 2018.
- Neural symbolic regression that scales. In International Conference on Machine Learning, pp. 936–945. Pmlr, 2021.
- Autonomous chemical research with large language models. Nature, 624(7992):570–578, 2023.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- ProxylessNAS: Direct neural architecture search on target task and hardware. In International Conference on Learning Representations, 2019.
- Large language models as tool makers. In International Conference on Learning Representations, 2024.
- Learning concise representations for regression by evolving networks of trees. In International Conference on Learning Representations, 2019.
- Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pp. 785–794, 2016.
- Chemberta: large-scale self-supervised pretraining for molecular property prediction. arXiv preprint arXiv:2010.09885, 2020.
- An overview of bilevel optimization. Annals of operations research, 153:235–256, 2007.
- Diffpd: Differentiable projective dynamics. ACM Transactions on Graphics (TOG), 41(2):1–21, 2021.
- Geometry-enhanced molecular representation learning for property prediction. Nature Machine Intelligence, 4(2):127–134, 2022.
- Science of science. Science, 359(6379):eaao0185, 2018.
- Halgren, T. A. Merck molecular force field. i. basis, form, scope, parameterization, and performance of mmff94. Journal of computational chemistry, 17(5-6):490–519, 1996.
- Hansen, N. The cma evolution strategy: a comparing review. Towards a new evolutionary computation: Advances in the estimation of distribution algorithms, pp. 75–102, 2006.
- Benchmarking large language models as ai research agents. arXiv preprint arXiv:2310.03302, 2023.
- Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024.
- The material point method for simulating continuum materials. In Acm siggraph 2016 courses, pp. 1–52. 2016.
- Junction tree variational autoencoder for molecular graph generation. In International conference on machine learning, pp. 2323–2332. PMLR, 2018.
- Bayesian symbolic regression. arXiv preprint arXiv:1910.08892, 2019.
- Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems, 30, 2017.
- Adam: A method for stochastic optimization. In International Conference on Learning Representations, San Diega, CA, USA, 2015.
- Parameter identification for symbolic regression using nonlinear least squares. Genetic Programming and Evolvable Machines, 21(3):471–501, 2020.
- Automated scientific discovery: From equation discovery to autonomous discovery systems. arXiv preprint arXiv:2305.02251, 2023.
- A probabilistic and multi-objective analysis of lexicase selection and ε𝜀\varepsilonitalic_ε-lexicase selection. Evolutionary Computation, 27(3):377–402, 2019.
- Contemporary symbolic regression methods and their relative performance. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, 2021.
- Landrum, G. et al. Rdkit: A software suite for cheminformatics, computational chemistry, and predictive modeling. Greg Landrum, 8:31, 2013.
- Empowering molecule discovery for molecule-caption translation with large language models: A chatgpt perspective. arXiv preprint arXiv:2306.06615, 2023.
- Transformer-based model for symbolic regression via joint supervised learning. In The Eleventh International Conference on Learning Representations, 2022.
- DARTS: Differentiable architecture search. In International Conference on Learning Representations, 2019.
- A general descent aggregation framework for gradient-based bi-level optimization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):38–57, 2022.
- Ai-based language models powering drug discovery and development. Drug Discovery Today, 26(11):2593–2607, 2021.
- Risp: Rendering-invariant state predictor with differentiable simulation and rendering for cross-domain parameter estimation. In International Conference on Learning Representations, 2021.
- Learning neural constitutive laws from motion observations for generalizable pde dynamics. In International Conference on Machine Learning. PMLR, 2023.
- Eureka: Human-level reward design via coding large language models. In International Conference on Learning Representations, 2024.
- Macklin, M. Warp: A high-performance python framework for gpu simulation and graphics, March 2022. NVIDIA GPU Technology Conference.
- McConaghy, T. Ffx: Fast, scalable, deterministic symbolic regression technology. Genetic Programming Theory and Practice IX, pp. 235–260, 2011.
- Symbolic regression via neural-guided genetic programming population seeding. In Advances in Neural Information Processing Systems, 2021.
- OpenAI. OpenAI: Introducing ChatGPT, 2022. URL https://openai.com/blog/chatgpt.
- OpenAI. OpenAI: GPT-4, 2023. URL https://openai.com/research/gpt-4.
- Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730–27744, 2022.
- Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Deep symbolic regression: Recovering mathematical expressions from data via risk-seeking policy gradients. In International Conference on Learning Representations, 2020.
- Popper, K. The logic of scientific discovery. Routledge, 2005.
- Quantum chemistry structures and properties of 134 kilo molecules. Scientific data, 1(1):1–7, 2014.
- Better informed distance geometry: using what we know to improve conformation generation. Journal of chemical information and modeling, 55(12):2562–2574, 2015.
- Mathematical discoveries from program search with large language models. Nature, pp. 1–3, 2023.
- Philosophy of science: A contemporary introduction. Routledge, 2019.
- Schapire, R. E. The boosting approach to machine learning: An overview. Nonlinear estimation and classification, pp. 149–171, 2003.
- Schneider, G. Automating drug discovery. Nature reviews drug discovery, 17(2):97–113, 2018.
- Chatgpt in drug discovery. 2023.
- Evolutionary algorithm for bilevel optimization using approximations of the lower level optimal solution mapping. European Journal of Operational Research, 257(2):395–411, 2017a.
- A review on bilevel optimization: From classical to evolutionary approaches and applications. IEEE Transactions on Evolutionary Computation, 22(2):276–295, 2017b.
- Application of a particle-in-cell method to solid mechanics. Computer physics communications, 87(1-2):236–252, 1995.
- Cognitive architectures for language agents. Transactions on Machine Learning Research, 2024. ISSN 2835-8856. Survey Certification.
- Solving olympiad geometry without human demonstrations. Nature, 625(7995):476–482, 2024.
- Ai feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity. Advances in Neural Information Processing Systems, 33:4860–4871, 2020.
- Symbolicgpt: A generative transformer model for symbolic regression. arXiv preprint arXiv:2106.14131, 2021.
- Linear scaling with and within semantic backpropagation-based genetic programming for symbolic regression. In Proceedings of the genetic and evolutionary computation conference, pp. 1084–1092, 2019.
- Improving model-based genetic programming for symbolic regression of small expressions. Evolutionary computation, 29(2):211–237, 2021.
- Scientific discovery in the age of artificial intelligence. Nature, 620(7972):47–60, 2023.
- Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
- Weininger, D. Smiles, a chemical language and information system. 1. introduction to methodology and encoding rules. Journal of chemical information and computer sciences, 28(1):31–36, 1988.
- A typology of scientific breakthroughs. Quantitative Science Studies, 1(3):1203–1222, 2020.
- Rethinking bi-level optimization in neural architecture search: A gibbs sampling perspective. In AAAI Conference on Artificial Intelligence, volume 35, pp. 10551–10559, 2021.
- Large language models as optimizers. In International Conference on Learning Representations, 2024.
- Tree of thoughts: Deliberate problem solving with large language models. In Conference on Neural Information Processing Systems, 2023a.
- ReAct: Synergizing reasoning and acting in language models. In International Conference on Learning Representations, 2023b.
- Population-based de novo molecule generation, using grammatical evolution. Chemistry Letters, 47(11):1431–1434, 2018.
- Efficiently programming large language models using sglang. arXiv preprint arXiv:2312.07104, 2023.
- Uni-mol: A universal 3d molecular representation learning framework. In International Conference on Learning Representations, 2023.
- Optimization of molecules via deep reinforcement learning. Scientific reports, 9(1):10752, 2019.
- Pingchuan Ma (91 papers)
- Tsun-Hsuan Wang (37 papers)
- Minghao Guo (45 papers)
- Zhiqing Sun (35 papers)
- Joshua B. Tenenbaum (257 papers)
- Daniela Rus (181 papers)
- Chuang Gan (195 papers)
- Wojciech Matusik (76 papers)