Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
120 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
3 tokens/sec
DeepSeek R1 via Azure Pro
55 tokens/sec
2000 character limit reached

BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments (2405.17631v3)

Published 27 May 2024 in cs.AI, cs.CE, and cs.MA

Abstract: Agents based on LLMs have shown great potential in accelerating scientific discovery by leveraging their rich background knowledge and reasoning capabilities. In this paper, we introduce BioDiscoveryAgent, an agent that designs new experiments, reasons about their outcomes, and efficiently navigates the hypothesis space to reach desired solutions. We demonstrate our agent on the problem of designing genetic perturbation experiments, where the aim is to find a small subset out of many possible genes that, when perturbed, result in a specific phenotype (e.g., cell growth). Utilizing its biological knowledge, BioDiscoveryAgent can uniquely design new experiments without the need to train a machine learning model or explicitly design an acquisition function as in Bayesian optimization. Moreover, BioDiscoveryAgent, using Claude 3.5 Sonnet, achieves an average of 21% improvement in predicting relevant genetic perturbations across six datasets, and a 46% improvement in the harder task of non-essential gene perturbation, compared to existing Bayesian optimization baselines specifically trained for this task. Our evaluation includes one dataset that is unpublished, ensuring it is not part of the LLM's training data. Additionally, BioDiscoveryAgent predicts gene combinations to perturb more than twice as accurately as a random baseline, a task so far not explored in the context of closed-loop experiment design. The agent also has access to tools for searching the biomedical literature, executing code to analyze biological datasets, and prompting another agent to critically evaluate its predictions. Overall, BioDiscoveryAgent is interpretable at every stage, representing an accessible new paradigm in the computational design of biological experiments with the potential to augment scientists' efficacy.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Significant-gravitas/auto-gpt: An experimental open-source attempt to make gpt-4 fully autonomous. https://github.com/Significant-Gravitas/Auto-GPT, 2023.
  2. Anthropic. Introducing claude, 2023. URL https://www.anthropic.com/index/introducing-claude.
  3. Autonomous chemical research with large language models. Nature, 624(7992):570–578, 2023.
  4. Rasa2 ablation in t cells boosts antigen sensitivity and long-term function. Nature, 609(7925):174–182, 2022.
  5. Y. T. Chen and J. Zou. Genept: A simple but hard-to-beat foundation model for genes and cells built from chatgpt. bioRxiv, 2023.
  6. Efficient generation of transcriptomic profiles by random composite measurements. Cell, 171(6):1424–1436, 2017.
  7. Empowering biomedical discovery with ai agents. arXiv preprint arXiv:2404.02831, 2024.
  8. Pooled screening of car t cells identifies diverse immune signaling domains for next-generation immunotherapies. Science translational medicine, 14(670):eabm1463, 2022.
  9. Mapping the genetic landscape of human cells. Cell, 174(4):953–967, 2018.
  10. W. Hou and Z. Ji. Assessing gpt-4 for cell type annotation in single-cell rna-seq analysis. Nature Methods, pages 1–4, 2024.
  11. Evaluation of large language models for discovery of gene set function. Research Square, 2023.
  12. Sequential optimal experimental design of perturbation screens guided by multi-modal priors. bioRxiv, pages 2023–12, 2023a.
  13. Benchmarking large language models as ai research agents. ArXiv, abs/2310.03302, 2023b. URL https://api.semanticscholar.org/CorpusID:263671541.
  14. T cells with chimeric antigen receptors have potent antitumor effects and can establish memory in patients with advanced leukemia. Science translational medicine, 3(95):95ra73–95ra73, 2011.
  15. Functional genomic hypothesis generation and experimentation by a robot scientist. Nature, 427(6971):247–252, 2004.
  16. Paperqa: Retrieval-augmented generative agent for scientific research. arXiv preprint arXiv:2312.07559, 2023.
  17. Scientific discovery: Computational explorations of the creative processes. 1987. URL https://api.semanticscholar.org/CorpusID:61752618.
  18. W. A. Lim. The emerging era of cell engineering: Harnessing the modularity of cells to program complex biological function. Science, 378(6622):848–852, 2022.
  19. Chatgpt-powered conversational drug editing using retrieval and domain feedback. arXiv preprint arXiv:2305.18090, 2023.
  20. Large language models to enhance bayesian optimization. arXiv preprint arXiv:2402.03921, 2024.
  21. Discobax discovery of optimal intervention sets in genomic experiment design. In International Conference on Machine Learning, pages 23170–23189. PMLR, 2023.
  22. Crispr screens decode cancer cell pathways that trigger γ𝛾\gammaitalic_γδ𝛿\deltaitalic_δ t cell detection. Nature, 621(7977):188–195, 2023.
  23. Genedisco: A benchmark for experimental design in drug discovery. In International Conference on Learning Representations, 2021.
  24. Med-flamingo: a multimodal medical few-shot learner. arXiv preprint arXiv:2307.15189, 2023.
  25. Going forward with genetics: recent technological advances and forward genetics in mice. The American journal of pathology, 182(5):1462–1473, 2013.
  26. The support of human genetic evidence for approved drug indications. Nature genetics, 47(8):856–860, 2015.
  27. Comparative performance evaluation of large language models for extracting molecular interactions and pathway knowledge. arXiv preprint arXiv:2307.08813, 2023a.
  28. Generative agents: Interactive simulacra of human behavior. arXiv preprint arXiv:2304.03442, 2023b.
  29. L. Przybyla and L. A. Gilbert. A new era in functional genomics screens. Nature Reviews Genetics, 23(2):89–103, 2022.
  30. Mathematical discoveries from program search with large language models. Nature, pages 1–3, 2023.
  31. Predicting transcriptional outcomes of novel multigene perturbations with gears. Nature Biotechnology, pages 1–9, 2023.
  32. Genome-wide crispr screen identifies protein pathways modulating tau protein levels in neurons. Communications biology, 4(1):736, 2021.
  33. Large language models are universal biomedical simulators. bioRxiv, pages 2023–06, 2023.
  34. An spns1-dependent lysosomal lipid transport pathway that enables cell survival under choline limitation. Science Advances, 9(16):eadf8966, 2023.
  35. Toolformer: Language models can teach themselves to use tools (2023). arXiv preprint arXiv:2302.04761, 2023.
  36. Toolformer: Language models can teach themselves to use tools. Advances in Neural Information Processing Systems, 36, 2024.
  37. Crispr activation and interference screens decode stimulation responses in primary human t cells. Science, 375(6580):eabj4008, 2022.
  38. K. Schneeberger. Using next-generation sequencing to isolate mutant genes from forward genetic screens. Nature Reviews Genetics, 15(10):662–676, 2014.
  39. N. Shakked and W. Zhang. Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381(6654):187–192, 2023. doi: 10.1126/science.adh2586. URL https://doi.org/10.1126/science.adh2586.
  40. Reflexion: an autonomous agent with dynamic memory and self-reflection. ArXiv, abs/2303.11366, 2023.
  41. Scientific discovery as problem solving. Synthese, 47:1–27, 1981. URL https://api.semanticscholar.org/CorpusID:46985700.
  42. Scientific discovery in the age of artificial intelligence. Nature, 620(7972):47–60, 2023.
  43. Crispr technology: A decade of genome editing is only the beginning. Science, 379(6629):eadd8643, 2023.
  44. Large language models are better reasoners with self-verification, 2023.
  45. G. Wobben. pymed - pubmed querying with python, 2020. URL https://github.com/gijswobben/pymed. Accessed: 03/2024.
  46. Zero-shot clinical trial patient matching with llms. arXiv preprint arXiv:2402.05125, 2024.
  47. React: Synergizing reasoning and acting in language models. ArXiv, abs/2210.03629, 2022.
Citations (3)

Summary

  • The paper demonstrates that BioDiscoveryAgent improves gene hit ratios by 18-29% through iterative experimental design using an LLM.
  • The agent integrates literature search, gene feature analysis, and an AI critic to deliver transparent and interpretable experimental decisions.
  • Experimental results validate the agent’s scalability and robustness, outperforming baseline methods by 130% in two-gene perturbations.

BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments

Introduction

The paper "BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments" presents a novel approach to accelerating the design of genetic perturbation experiments using a specialized AI agent, BioDiscoveryAgent. This AI agent leverages a LLM to generate hypotheses, plan experiments, and interpret results with the ultimate goal of identifying genes that, when perturbed, result in a specific phenotype.

Methodology

BioDiscoveryAgent addresses the inefficiencies and complexities involved in traditional genetic perturbation experiment designs that employ bespoke machine learning models and acquisition functions. The agent integrates domain-specific knowledge, interpretable decision-making, and a suite of external tools, thus creating a robust framework for experimental design.

The core of BioDiscoveryAgent is powered by an LLM. Unlike conventional machine learning approaches requiring specific training, the LLM within BioDiscoveryAgent leverages pre-existing biological knowledge and reasoning capabilities. The agent functions in a closed-loop setup, iteratively designing experiments based on the results from previous ones. It operates through a structured prompt-response mechanism, processing experimental observations and forming hypotheses for subsequent rounds of gene perturbations.

Tools and Features

BioDiscoveryAgent incorporates multiple tools to enhance its performance:

  1. Literature Search: The agent can query biomedical literature databases and summarize findings to inform its gene selection. This feature adds verifiable citations to its decisions and enriches its knowledge base.
  2. Gene Search Based on Features: By evaluating the similarity or dissimilarity of gene features obtained from tabular datasets, the agent can identify genes that may not be well-represented in the textual data the LLM was trained on.
  3. AI Critic: A secondary LLM acts as a critic, evaluating and refining the primary agent's gene selection. This iterative feedback loop improves decision-making quality and gene selection diversity.

Experimental Results

BioDiscoveryAgent’s performance was validated using data from multiple genetic perturbation experiments across different contexts and cell types. The evaluation employed a comparative analysis against baseline methods that utilize conventional machine learning models with various acquisition functions.

Numerical Results:

  • In single-gene perturbation experiments, BioDiscoveryAgent demonstrated an average of 18% improvement in hit ratio over baseline methods across five datasets.
  • For non-essential genes, which represent a more challenging subset, the agent reported a 29% higher hit detection rate than baselines, underscoring its capability to predict challenging yet biologically significant hits.
  • In the two-gene perturbation task, the agent outperformed the random sampling baseline by 130%.

Robustness and Interpretability

The paper emphasizes the interpretability of BioDiscoveryAgent’s decision-making process. Each stage of the agent's reasoning, from hypothesis formation to gene selection, is transparent and can be traced back to specific literature references or experimental observations. This traceability is crucial for practical applications where experimental results must be reproducible and scientifically valid.

Future Implications

BioDiscoveryAgent’s architecture and functionality suggest significant implications for future developments in artificial intelligence and biological research:

  1. Scalability: The framework is scalable to other types of high-dimensional biological data and can be adapted for various experimental contexts beyond gene perturbation.
  2. Human-AI Collaboration: By providing an interpretable and efficient experimental design process, the agent facilitates closer collaboration between human researchers and AI, potentially shifting the paradigm of how biological experiments are conceived and conducted.
  3. Generalizability: The method can be generalized to other domains requiring iterative experimentation and complex hypothesis testing, thus broadening the impact of LLMs in scientific discovery.

Conclusion

The paper showcases BioDiscoveryAgent as an effective and interpretable AI agent for designing genetic perturbation experiments. By integrating LLMs with specialized tools for literature search, gene feature analysis, and AI-based critique, the agent overcomes the limitations of traditional machine learning-based experimental designs. This innovative approach significantly enhances the efficiency and accuracy of identifying genes that influence specific phenotypes, opening new avenues for biological research and interdisciplinary applications of AI.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com