Monte Carlo Thought Search: Large Language Model Querying for Complex Scientific Reasoning in Catalyst Design (2310.14420v1)
Abstract: Discovering novel catalysts requires complex reasoning involving multiple chemical properties and resultant trade-offs, leading to a combinatorial growth in the search space. While LLMs (LLM) have demonstrated novel capabilities for chemistry through complex instruction following capabilities and high quality reasoning, a goal-driven combinatorial search using LLMs has not been explored in detail. In this work, we present a Monte Carlo Tree Search-based approach that improves beyond state-of-the-art chain-of-thought prompting variants to augment scientific reasoning. We introduce two new reasoning datasets: 1) a curation of computational chemistry simulations, and 2) diverse questions written by catalysis researchers for reasoning about novel chemical conversion processes. We improve over the best baseline by 25.8\% and find that our approach can augment scientist's reasoning and discovery process with novel insights.
- Sustainable conversion of carbon dioxide: an integrated review of catalysis and life cycle assessment. Chemical reviews, 118(2):434–504.
- Emergent autonomous scientific research capabilities of large language models. arXiv preprint arXiv:2304.05332.
- Chemcrow: Augmenting large-language models with chemistry tools. arXiv preprint arXiv:2304.05376.
- Mustafa Canakci and Jon Van Gerpen. 1999. Biodiesel production viaacid catalysis. Transactions of the ASAE, 42(5):1203–1210.
- Cayque Monteiro Castro Nascimento and André Silva Pimentel. 2023. Do large language models understand chemistry? a conversation with chatgpt. Journal of Chemical Information and Modeling, 63(6):1649–1655.
- The open catalyst 2020 (oc20) dataset and community challenges. arxiv. arXiv.
- Shuan Chen and Yousung Jung. 2022. A generalized-template-based graph neural network for accurate organic reactivity prediction. Nature Machine Intelligence, 4(9):772–780.
- Group selfies: a robust fragment-based molecular string representation. Digital Discovery.
- Chemberta: Large-scale self-supervised pretraining for molecular property prediction. arXiv preprint arXiv:2010.09885.
- Unifying molecular and textual representations via multi-task language modelling. arXiv preprint arXiv:2301.12586.
- Antonia Creswell and Murray Shanahan. 2022. Faithful reasoning using large language models. arXiv preprint arXiv:2208.14271.
- Selection-inference: Exploiting large language models for interpretable logical reasoning. arXiv preprint arXiv:2205.09712.
- Yolanda A Daza and John N Kuhn. 2016. Co 2 conversion by reverse water gas shift catalysis: comparison of catalysts, mechanisms and their consequences for co 2 conversion to liquid fuels. RSC advances, 6(55):49675–49691.
- RLPrompt: Optimizing discrete text prompts with reinforcement learning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 3369–3391, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Language model cascades. arXiv preprint arXiv:2207.10342.
- Translation between molecules and natural language. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 375–413, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Text2Mol: Cross-modal molecule retrieval with natural language queries. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 595–607, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Molecular representation learning with language models and domain-relevant auxiliary tasks. arXiv preprint arXiv:2011.13230.
- Gemnet: Universal directional graph neural networks for molecules. Advances in Neural Information Processing Systems, 34:6790–6802.
- Glen M Hocky and Andrew D White. 2022. Natural language processing models that automate programming will transform chemistry research and teaching. Digital discovery, 1(2):79–83.
- Is gpt-3 all you need for low-data discovery in chemistry? ChemRxiv preprint.
- Junction tree variational autoencoder for molecular graph generation. In International conference on machine learning, pages 2323–2332. PMLR.
- Maieutic prompting: Logically consistent reasoning with recursive explanations. arXiv preprint arXiv:2205.11822.
- Tuning selectivity of co2 hydrogenation reactions at the metal/oxide interface. Journal of the American Chemical Society, 139(29):9739–9754.
- Large language models are zero-shot reasoners. arXiv preprint arXiv:2205.11916.
- Self-referencing embedded strings (selfies): A 100% robust molecular string representation. Machine Learning: Science and Technology, 1(4):045024.
- Knowledge-enhanced biomedical language models. In Journal of Biomedical Informatics.
- Multi-modal molecule structure-text model for text-based retrieval and editing. arXiv preprint arXiv:2212.10789.
- Chatgpt-powered conversational drug editing using retrieval and domain feedback. arXiv preprint arXiv:2305.18090.
- A text-guided protein design framework. arXiv preprint arXiv:2302.04611.
- Molxpt: Wrapping molecules with text for generative pre-training. arXiv preprint arXiv:2305.10688.
- Current status and challenges in the heterogeneous catalysis for biodiesel production. Renewable and Sustainable Energy Reviews, 157:112012.
- Density functional theory in surface chemistry and catalysis. Proceedings of the National Academy of Sciences, 108(3):937–943.
- NVIDIA Corporation. 2022. Megamolbart v0.2.
- OpenAI. 2023. Gpt-4 technical report.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- Bayesian optimization of catalysts with in-context learning. arXiv preprint arXiv:2304.05341.
- Schnet–a deep learning architecture for molecules and materials. The Journal of Chemical Physics, 148(24):241722.
- Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction. ACS central science, 5(9):1572–1583.
- Mapping the space of chemical reactions using attention-based neural networks. Nature Machine Intelligence, 3(2):144–152.
- Enhancing activity prediction models in drug discovery with the ability to understand human language. arXiv preprint arXiv:2303.03363.
- Challenges and prospects in solar water splitting and co2 reduction with inorganic and hybrid nanostructures. ACS Catalysis, 8(4):3602–3635.
- Current and future roles of artificial intelligence in medicinal chemistry synthesis. Journal of medicinal chemistry, 63(16):8667–8682.
- A molecular multimodal foundation model associating molecule graphs with natural language. arXiv preprint arXiv:2209.05481.
- Galactica: A large language model for science. arXiv preprint arXiv:2211.09085.
- The open catalyst 2022 (oc22) dataset and challenges for oxide electrocatalysts. ACS Catalysis, 13(5):3066–3084.
- Can we quickly learn to “translate” bioactive molecules with transformer models? Journal of Chemical Information and Modeling, 63(6):1734–1744.
- Dual use of artificial-intelligence-powered drug discovery. Nature Machine Intelligence, 4(3):189–191.
- Bioassayclr: Prediction of biological activity for novel bioassays based on rich textual descriptions. In ELLIS ML4Molecules workshop.
- Inferring experimental procedures from text-based representations of chemical reactions. Nature communications, 12(1):2573.
- Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171.
- Benchmarking deep graph generative models for optimizing new drug molecules for covid-19. arXiv preprint arXiv:2102.04977.
- Chain of thought prompting elicits reasoning in large language models. arXiv preprint arXiv:2201.11903.
- David Weininger. 1988. Smiles, a chemical language and information system. 1. introduction to methodology and encoding rules. Journal of chemical information and computer sciences, 28(1):31–36.
- Smiles. 2. algorithm for generation of unique smiles notation. Journal of chemical information and computer sciences, 29(2):97–101.
- Do large language models know chemistry? ChemRxiv preprint.
- Assessment of chemistry knowledge in large language models that generate code. Digital Discovery, 2(2):368–376.
- Hanwen Xu and Sheng Wang. 2022. Protranslator: zero-shot protein function prediction using textual description. In Research in Computational Molecular Biology: 26th Annual International Conference, RECOMB 2022, San Diego, CA, USA, May 22–25, 2022, Proceedings, pages 279–294. Springer.
- Protst: Multi-modality learning of protein sequences and biomedical texts. arXiv preprint arXiv:2301.12040.
- Shenzhen Xu and Emily A Carter. 2018. Theoretical insights into heterogeneous (photo) electrochemical co2 reduction. Chemical reviews, 119(11):6631–6669.
- Machine learning in catalysis, from proposal to practicing. ACS omega, 5(1):83–88.
- Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
- Graph convolutional policy network for goal-directed molecular graph generation. Advances in neural information processing systems, 31.
- A deep-learning system bridging molecule structure and biomedical text with comprehension comparable to human professionals. Nature communications, 13(1):862.
- Adversarial modality alignment network for cross-modal molecule retrieval. IEEE Transactions on Artificial Intelligence.
- An introduction to electrocatalyst design using machine learning for renewable energy storage. arXiv preprint arXiv:2010.09435.