DPA-2: a large atomic model as a multi-task learner (2312.15492v2)
Abstract: The rapid advancements in AI are catalyzing transformative changes in atomic modeling, simulation, and design. AI-driven potential energy models have demonstrated the capability to conduct large-scale, long-duration simulations with the accuracy of ab initio electronic structure methods. However, the model generation process remains a bottleneck for large-scale applications. We propose a shift towards a model-centric ecosystem, wherein a large atomic model (LAM), pre-trained across multiple disciplines, can be efficiently fine-tuned and distilled for various downstream tasks, thereby establishing a new framework for molecular modeling. In this study, we introduce the DPA-2 architecture as a prototype for LAMs. Pre-trained on a diverse array of chemical and materials systems using a multi-task approach, DPA-2 demonstrates superior generalization capabilities across multiple downstream tasks compared to the traditional single-task pre-training and fine-tuning methodologies. Our approach sets the stage for the development and broad application of LAMs in molecular and materials simulation research.
- Inhomogeneous electron gas. Physical review, 136(3B):B864, 1964.
- Self-consistent equations including exchange and correlation effects. Physical review, 140(4A):A1133, 1965.
- Combined free-energy calculation and machine learning methods for understanding ligand unbinding kinetics. Journal of chemical theory and computation, 18(4):2543–2555, 2022.
- Qdπ𝜋\piitalic_π: A quantum deep potential interaction model for drug discovery. Journal of Chemical Theory and Computation, 19(4):1261–1275, 2023.
- Machine learning a general-purpose interatomic potential for silicon. Physical Review X, 8(4):041048, 2018.
- A general-purpose machine-learning force field for bulk and nanostructured phosphorus. Nature communications, 11(1):1–11, 2020.
- Deep potentials for materials science. Materials Futures, 1(2):022601, 2022.
- Machine learning for atomic simulation and activity prediction in heterogeneous catalysis: current status and future. ACS Catalysis, 10(22):13213–13226, 2020.
- Ammonia decomposition on lithium imide surfaces: A new paradigm in heterogeneous catalysis. ChemRxiv, 2022.
- Unified approach for molecular dynamics and density-functional theory. Physical review letters, 55(22):2471, 1985.
- Dp-gen: A concurrent learning platform for the generation of reliable deep learning based potential energy models. Computer Physics Communications, 253:107206, 2020.
- Less is more: Sampling chemical space with active learning. The Journal of chemical physics, 148(24):241733, 2018.
- Active learning in gaussian process interpolation of potential energy surfaces. The Journal of chemical physics, 149(17):174114, 2018.
- Accurate deep potential model for the al–cu–mg alloy in the full concentration space. Chinese Physics B, 30(5):050706, 2021.
- A universal graph deep learning interatomic potential for the periodic table. arXiv preprint arXiv:2202.02450, 2022.
- Towards universal neural network potential for material discovery applicable to arbitrary combination of 45 elements. Nature Communications, 13(1):2991, 2022.
- Unified graph neural network force-field for the periodic table: solid state applications. Digital Discovery, 2(2):346–355, 2023.
- Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
- Gpip: Geometry-enhanced pre-training on interatomic potentials. arXiv preprint arXiv:2309.15718, 2023.
- May the force be with you: Unified force-centric pre-training for 3d molecular conformations. arXiv preprint arXiv:2308.14759, 2023.
- Leveraging multitask learning to improve the transferability of machine learned force fields. ChemRxiv, 2023.
- Denoise pre-training on non-equilibriummolecules for accurate and transferable neuralpotentials. arXiv preprint arXiv:2303.02216, 2023.
- Iasonas Kokkinos. Ubernet: Training a universal convolutional neural network for low-, mid-, and high-level vision using diverse datasets and limited memory. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6129–6138, 2017.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- From molecules to materials: Pre-training large generalizable models for atomic property prediction. arXiv preprint arXiv:2310.16802, 2023.
- Approaching coupled cluster accuracy with a general-purpose neural network potential through transfer learning. Nature communications, 10(1):1–8, 2019.
- Transfer learning using attentions across atomic systems with graph neural networks (taag). The Journal of Chemical Physics, 156(18):184702, 2022.
- Deep learning. MIT press, 2016.
- Mark Tuckerman. Statistical Mechanics: Theory and Molecular Simulation. OUP Oxford, 2010.
- How do graph networks generalize to large and diverse molecular systems? arXiv preprint arXiv:2204.02782, 2022.
- Equiformerv2: Improved equivariant transformer for scaling to higher-degree representations. arXiv preprint arXiv:2306.12059, 2023.
- E (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature communications, 13(1):1–11, 2022.
- Learning local equivariant representations for large-scale atomistic dynamics. arXiv preprint arXiv:2204.05249, 2022.
- Generalized neural-network representation of high-dimensional potential-energy surfaces. Physical review letters, 98(14):146401, 2007.
- Ani-1: an extensible neural network potential with dft accuracy at force field computational cost. Chemical science, 8(4):3192–3203, 2017.
- Quantum-chemical insights from deep tensor neural networks. Nature communications, 8(1):1–8, 2017.
- wacsf—weighted atom-centered symmetry functions as descriptors in machine learning potentials. The Journal of chemical physics, 148(24), 2018.
- Deep potential molecular dynamics: a scalable model with the accuracy of quantum mechanics. Physical review letters, 120(14):143001, 2018.
- End-to-end symmetry preserving inter-atomic potential energy model for finite and extended systems. Advances in Neural Information Processing Systems, 31, 2018.
- Deepmd-kit v2: A software package for deep potential models. The Journal of Chemical Physics, 159:054801, 2023.
- Dpa-1: Pretraining of attention-based deep potential model for molecular simulation. arXiv preprint arXiv:2208.08236, 2022.
- Embedded atom neural network potentials: Efficient and accurate machine learning with a physically inspired representation. The journal of physical chemistry letters, 10(17):4962–4967, 2019.
- Gaussian approximation potentials: The accuracy of quantum mechanics, without the electrons. Physical review letters, 104(13):136403, 2010.
- On representing chemical environments. Physical Review B, 87(18):184115, 2013.
- Fast and accurate modeling of molecular atomization energies with machine learning. Physical review letters, 108(5):058301, 2012.
- Machine learning of accurate energy-conserving molecular force fields. Science advances, 3(5):e1603015, 2017.
- Spectral neighbor analysis method for automated generation of quantum-accurate interatomic potentials. Journal of Computational Physics, 285:316–330, 2015.
- Alexander V Shapeev. Moment tensor potentials: A class of systematically improvable interatomic potentials. Multiscale Modeling & Simulation, 14(3):1153–1173, 2016.
- Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds. arXiv preprint arXiv:1802.08219, 2018.
- Se (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. arXiv preprint arXiv:2101.03164, 2021.
- Schnet: A continuous-filter convolutional neural network for modeling quantum interactions. Advances in neural information processing systems, 30, 2017.
- Fast and uncertainty-aware directional message passing for non-equilibrium molecules. arXiv preprint arXiv:2011.14115, 2020.
- Equivariant message passing for the prediction of tensorial properties and molecular spectra. arXiv preprint arXiv:2102.03150, 2021.
- Gemnet: Universal directional graph neural networks for molecules. Advances in Neural Information Processing Systems, 34:6790–6802, 2021.
- Rotation invariant graph neural networks using spin convolutions. arXiv preprint arXiv:2106.09575, 2021.
- Spherical channels for modeling atomic interactions. Advances in Neural Information Processing Systems, 35:8054–8067, 2022.
- Equiformer: Equivariant graph attention transformer for 3d atomistic graphs. arXiv preprint arXiv:2206.11990, 2022.
- Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8):1798–1828, 2013.
- Network representation learning: A survey. IEEE transactions on Big Data, 6(1):3–28, 2018.
- Improving language understanding by generative pre-training. 2018.
- An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
- Uni-mol: A universal 3d molecular representation learning framework. ChemRxiv, 2022.
- Unified 2d and 3d pre-training of molecular representations. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 2626–2636, 2022.
- Pre-training via denoisingfor molecular property prediction. arXiv preprint arXiv:2206.00133, 2022.
- Highly accurate quantum chemical propertyprediction with uni-mol+. arXiv preprint arXiv:2303.16982, 2023.
- Fractional denoising for 3d molecular pre-training. arXiv preprint arXiv:2307.10683, 2023.
- 3d equivariant molecular graph pretraining. arXiv preprint arXiv:2207.08824, 2022.
- D. Beaini and et.al. Towards foundational models for molecularlearning on large-scale multi-task datasets. arXiv:2310.04292, 2023.
- K.L.K. Lee and et.al. Towards foundation models for materials science: The open matsci ml toolkit. arXiv:2310.07864, 2023.
- Synthetic pre-training for neural-networkinteratomic potentials. arXiv preprint arXiv:2307.15714, 2023.
- Open catalyst 2020 (oc20) dataset and community challenges. ACS Catalysis, 11(10):6059–6072, 2021.
- Dp compress: A model compression scheme for generating efficient deep potential models. Journal of chemical theory and computation, 18(9):5559–5567, 2022.
- Efficiency of ab-initio total energy calculations for metals and semiconductors using a plane-wave basis set. Computational materials science, 6(1):15–50, 1996.
- Efficient iterative schemes for ab initio total-energy calculations using a plane-wave basis set. Physical review B, 54(16):11169, 1996.
- Gaussian 16, 2016.
- Systematically improvable optimized atomic basis sets for ab initio calculations. Journal of Physics: Condensed Matter, 22(44):445501, 2010.
- Large-scale ab initio simulations based on systematically improvable atomic basis. Computational Materials Science, 112:503–517, 2016.
- Gemnet-oc: developing graph neural networks for large and diverse molecular simulation datasets. arXiv preprint arXiv:2204.02782, 2022.
- The individual and collective effects of exact exchange and dispersion interactions on the ab initio structure of liquid water. The Journal of chemical physics, 141(8), 2014.
- Deep potential generation scheme and simulation protocol for the li10gep2s12-type superionic conductors. The Journal of Chemical Physics, 154(9):094703, 2021.
- First principles study of the li10gep2s12 lithium super ionic conductor material. Chemistry of Materials, 24(1):15–17, 2012.
- Ionic correlations and failure of nernst-einstein relation in solid-state electrolytes. Physical Review Materials, 1(2):025402, 2017.
- Single-crystal x-ray structure analysis of the superionic conductor li 10 gep 2 s 12. Physical Chemistry Chemical Physics, 15(28):11620–11622, 2013.
- Universal interatomic potential for perovskite oxides. Physical Review B, 108(18):L180104, 2023.
- Phase diagram and properties of pb (in1/2nb1/2) o3–pb (mg1/3nb2/3) o3–pbtio3 polycrystalline ceramics. Journal of the European Ceramic Society, 32(2):433–439, 2012.
- Soft phonon modes and diffuse scattering in pb (in1/2nb1/2) o3-pb (mg1/3nb2/3) o3-pbtio3 relaxor ferroelectrics. Journal of Materiomics, 4(4):345–352, 2018.
- Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Generalized gradient approximation made simple. Physical review letters, 77(18):3865, 1996.
- Cp2k: An electronic structure and molecular dynamics software package-quickstep: Efficient and accurate electronic structure calculations. The Journal of Chemical Physics, 152(19):194103, 2020.
- A consistent and accurate ab initio parametrization of density functional dispersion correction (dft-d) for the 94 elements h-pu. The Journal of chemical physics, 132(15), 2010.
- The ebi rdf platform: linked open data for the life sciences. Bioinformatics, 30(9):1338–1339, 2014.
- The chembl bioactivity database: an update. Nucleic acids research, 42(D1):D1083–D1090, 2014.
- Mychembl: a virtual platform for distributing cheminformatics tools and open data. Challenges, 5(2):334–337, 2014.
- Machine-learning-based interatomic potentials for group iib to via semiconductors: A comparative study of universal and independent models. arXiv preprint arXiv:2311.11305, 2023.
- Phase diagram of a deep potential water model. Physical review letters, 126(23):236001, 2021.
- A generalizable machine learning potential of ag–au nanoalloys and its application to surface reconstruction, segregation and diffusion. Modelling and Simulation in Materials Science and Engineering, 30(2):025003, 2021.
- Virtual exploration of the small-molecule chemical universe below 160 daltons. Angewandte Chemie International Edition, 44(10):1504–1508, 2005.
- Virtual exploration of the chemical universe up to 11 atoms of c, n, o, f: assembly of 26.4 million structures (110.9 million stereoisomers) and analysis for new ring systems, stereochemistry, physicochemical properties, compound classes, and drug discovery. Journal of chemical information and modeling, 47(2):342–353, 2007.
- 970 million druglike small molecules for virtual screening in the chemical universe database gdb-13. Journal of the American Chemical Society, 131(25):8732–8733, 2009.
- Transition1x-a dataset for building generalizable reactive machine learning potentials. Scientific Data, 9(1):779, 2022.
- Enumeration of 166 billion organic small molecules in the chemical universe database gdb-17. Journal of chemical information and modeling, 52(11):2864–2875, 2012.
- Modeling the high-pressure solid and liquid phases of tin from deep potentials with ab initio accuracy. Physical Review Materials, 7(5):053603, 2023.
- Specialising neural network potentials for accurate properties and application to the mechanical response of titanium. npj Computational Materials, 7(1):206, 2021.
- Classical and machine learning interatomic potentials for bcc vanadium. Physical Review Materials, 6(11):113603, 2022.
- A tungsten deep neural-network potential for simulating mechanical property degradation under fusion service environment. Nuclear Fusion, 2022.
- Exploring the chemical space of linear alkane pyrolysis via deep potential generator. Energy & fuels, 35(1):762–769, 2020.
- Deep learning of accurate force field of ferroelectric hfo 2. Physical Review B, 103(2):024108, 2021.
- Accurate force field of two-dimensional ferroelectrics from deep learning. Physical Review B, 104(17):174107, 2021.
- A deep potential model with long-range electrostatic interactions. The Journal of Chemical Physics, 156(12), 2022.
- Modeling liquid water by climbing up jacob’s ladder in density functional theory facilitated by using deep neural network potentials. The Journal of Physical Chemistry B, 125(41):11444–11456, 2021.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.