BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments (2405.17631v3)
Abstract: Agents based on LLMs have shown great potential in accelerating scientific discovery by leveraging their rich background knowledge and reasoning capabilities. In this paper, we introduce BioDiscoveryAgent, an agent that designs new experiments, reasons about their outcomes, and efficiently navigates the hypothesis space to reach desired solutions. We demonstrate our agent on the problem of designing genetic perturbation experiments, where the aim is to find a small subset out of many possible genes that, when perturbed, result in a specific phenotype (e.g., cell growth). Utilizing its biological knowledge, BioDiscoveryAgent can uniquely design new experiments without the need to train a machine learning model or explicitly design an acquisition function as in Bayesian optimization. Moreover, BioDiscoveryAgent, using Claude 3.5 Sonnet, achieves an average of 21% improvement in predicting relevant genetic perturbations across six datasets, and a 46% improvement in the harder task of non-essential gene perturbation, compared to existing Bayesian optimization baselines specifically trained for this task. Our evaluation includes one dataset that is unpublished, ensuring it is not part of the LLM's training data. Additionally, BioDiscoveryAgent predicts gene combinations to perturb more than twice as accurately as a random baseline, a task so far not explored in the context of closed-loop experiment design. The agent also has access to tools for searching the biomedical literature, executing code to analyze biological datasets, and prompting another agent to critically evaluate its predictions. Overall, BioDiscoveryAgent is interpretable at every stage, representing an accessible new paradigm in the computational design of biological experiments with the potential to augment scientists' efficacy.
- Significant-gravitas/auto-gpt: An experimental open-source attempt to make gpt-4 fully autonomous. https://github.com/Significant-Gravitas/Auto-GPT, 2023.
- Anthropic. Introducing claude, 2023. URL https://www.anthropic.com/index/introducing-claude.
- Autonomous chemical research with large language models. Nature, 624(7992):570–578, 2023.
- Rasa2 ablation in t cells boosts antigen sensitivity and long-term function. Nature, 609(7925):174–182, 2022.
- Y. T. Chen and J. Zou. Genept: A simple but hard-to-beat foundation model for genes and cells built from chatgpt. bioRxiv, 2023.
- Efficient generation of transcriptomic profiles by random composite measurements. Cell, 171(6):1424–1436, 2017.
- Empowering biomedical discovery with ai agents. arXiv preprint arXiv:2404.02831, 2024.
- Pooled screening of car t cells identifies diverse immune signaling domains for next-generation immunotherapies. Science translational medicine, 14(670):eabm1463, 2022.
- Mapping the genetic landscape of human cells. Cell, 174(4):953–967, 2018.
- W. Hou and Z. Ji. Assessing gpt-4 for cell type annotation in single-cell rna-seq analysis. Nature Methods, pages 1–4, 2024.
- Evaluation of large language models for discovery of gene set function. Research Square, 2023.
- Sequential optimal experimental design of perturbation screens guided by multi-modal priors. bioRxiv, pages 2023–12, 2023a.
- Benchmarking large language models as ai research agents. ArXiv, abs/2310.03302, 2023b. URL https://api.semanticscholar.org/CorpusID:263671541.
- T cells with chimeric antigen receptors have potent antitumor effects and can establish memory in patients with advanced leukemia. Science translational medicine, 3(95):95ra73–95ra73, 2011.
- Functional genomic hypothesis generation and experimentation by a robot scientist. Nature, 427(6971):247–252, 2004.
- Paperqa: Retrieval-augmented generative agent for scientific research. arXiv preprint arXiv:2312.07559, 2023.
- Scientific discovery: Computational explorations of the creative processes. 1987. URL https://api.semanticscholar.org/CorpusID:61752618.
- W. A. Lim. The emerging era of cell engineering: Harnessing the modularity of cells to program complex biological function. Science, 378(6622):848–852, 2022.
- Chatgpt-powered conversational drug editing using retrieval and domain feedback. arXiv preprint arXiv:2305.18090, 2023.
- Large language models to enhance bayesian optimization. arXiv preprint arXiv:2402.03921, 2024.
- Discobax discovery of optimal intervention sets in genomic experiment design. In International Conference on Machine Learning, pages 23170–23189. PMLR, 2023.
- Crispr screens decode cancer cell pathways that trigger γ𝛾\gammaitalic_γδ𝛿\deltaitalic_δ t cell detection. Nature, 621(7977):188–195, 2023.
- Genedisco: A benchmark for experimental design in drug discovery. In International Conference on Learning Representations, 2021.
- Med-flamingo: a multimodal medical few-shot learner. arXiv preprint arXiv:2307.15189, 2023.
- Going forward with genetics: recent technological advances and forward genetics in mice. The American journal of pathology, 182(5):1462–1473, 2013.
- The support of human genetic evidence for approved drug indications. Nature genetics, 47(8):856–860, 2015.
- Comparative performance evaluation of large language models for extracting molecular interactions and pathway knowledge. arXiv preprint arXiv:2307.08813, 2023a.
- Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442, 2023b.
- L. Przybyla and L. A. Gilbert. A new era in functional genomics screens. Nature Reviews Genetics, 23(2):89–103, 2022.
- Mathematical discoveries from program search with large language models. Nature, pages 1–3, 2023.
- Predicting transcriptional outcomes of novel multigene perturbations with gears. Nature Biotechnology, pages 1–9, 2023.
- Genome-wide crispr screen identifies protein pathways modulating tau protein levels in neurons. Communications biology, 4(1):736, 2021.
- Large language models are universal biomedical simulators. bioRxiv, pages 2023–06, 2023.
- An spns1-dependent lysosomal lipid transport pathway that enables cell survival under choline limitation. Science Advances, 9(16):eadf8966, 2023.
- Toolformer: Language models can teach themselves to use tools (2023). arXiv preprint arXiv:2302.04761, 2023.
- Toolformer: Language models can teach themselves to use tools. Advances in Neural Information Processing Systems, 36, 2024.
- Crispr activation and interference screens decode stimulation responses in primary human t cells. Science, 375(6580):eabj4008, 2022.
- K. Schneeberger. Using next-generation sequencing to isolate mutant genes from forward genetic screens. Nature Reviews Genetics, 15(10):662–676, 2014.
- N. Shakked and W. Zhang. Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381(6654):187–192, 2023. doi: 10.1126/science.adh2586. URL https://doi.org/10.1126/science.adh2586.
- Reflexion: an autonomous agent with dynamic memory and self-reflection. ArXiv, abs/2303.11366, 2023.
- Scientific discovery as problem solving. Synthese, 47:1–27, 1981. URL https://api.semanticscholar.org/CorpusID:46985700.
- Scientific discovery in the age of artificial intelligence. Nature, 620(7972):47–60, 2023.
- Crispr technology: A decade of genome editing is only the beginning. Science, 379(6629):eadd8643, 2023.
- Large language models are better reasoners with self-verification, 2023.
- G. Wobben. pymed - pubmed querying with python, 2020. URL https://github.com/gijswobben/pymed. Accessed: 03/2024.
- Zero-shot clinical trial patient matching with llms. arXiv preprint arXiv:2402.05125, 2024.
- React: Synergizing reasoning and acting in language models. ArXiv, abs/2210.03629, 2022.