Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Language models can generate molecules, materials, and protein binding sites directly in three dimensions as XYZ, CIF, and PDB files (2305.05708v1)

Published 9 May 2023 in cs.LG and q-bio.QM

Abstract: LLMs are powerful tools for molecular design. Currently, the dominant paradigm is to parse molecular graphs into linear string representations that can easily be trained on. This approach has been very successful, however, it is limited to chemical structures that can be completely represented by a graph -- like organic molecules -- while materials and biomolecular structures like protein binding sites require a more complete representation that includes the relative positioning of their atoms in space. In this work, we show how LLMs, without any architecture modifications, trained using next-token prediction -- can generate novel and valid structures in three dimensions from various substantially different distributions of chemical structures. In particular, we demonstrate that LLMs trained directly on sequences derived directly from chemical file formats like XYZ files, Crystallographic Information files (CIFs), or Protein Data Bank files (PDBs) can directly generate molecules, crystals, and protein binding sites in three dimensions. Furthermore, despite being trained on chemical file sequences -- LLMs still achieve performance comparable to state-of-the-art models that use graph and graph-derived string representations, as well as other domain-specific 3D generative models. In doing so, we demonstrate that it is not necessary to use simplified molecular representations to train chemical LLMs -- that they are powerful generative models capable of directly exploring chemical space in three dimensions for very different structures.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Daniel Flam-Shepherd (9 papers)
  2. Alán Aspuru-Guzik (227 papers)
Citations (41)
X Twitter Logo Streamline Icon: https://streamlinehq.com