Machine-Learned Interatomic Potentials
- Machine-learned interatomic potentials are data-driven models that predict atomic energies and forces with near-quantum accuracy using locality-based energy partitioning.
- They leverage symmetry-preserving descriptors like SOAP and Behler–Parrinello functions with regression models including neural networks and kernel methods for enhanced transferability.
- Applications span molecular dynamics, defect analysis, and phase transitions, employing active learning and multiscale strategies to optimize accuracy and computational cost.
Machine-learned interatomic potentials (MLIPs) are data-driven models that predict atomic energies and forces with near-first-principles accuracy and are employed as surrogates for explicit quantum-mechanical calculations in atomistic simulations. MLIPs have rapidly supplanted classical analytic potentials in many areas of materials science, computational chemistry, and molecular simulation, offering systematic improvements in accuracy, transferability, and flexibility. Their deployment now spans static structure optimization, molecular dynamics (MD) of complex and anharmonic systems, high-throughput property screening, and multi-scale modeling frameworks.
1. Mathematical Foundations and Model Architectures
MLIPs universally adopt a locality-based energy partitioning, expressing the total energy as a sum over atomic contributions: where each depends only on neighboring atoms within cutoff radius (Mishin, 2021). To achieve rotational, translational, and permutational invariance, diverse descriptors and regression models have been designed:
Descriptors:
- Behler–Parrinello symmetry functions (BPNN): Radial and angular environment functions for each atom (Mishin, 2021, Rili, 2024).
- SOAP: "Smooth Overlap of Atomic Positions" expands neighbor densities in a Hilbert space using spherical harmonics and radial basis functions, yielding a rotationally invariant power spectrum (Mishin, 2021, Robredo-Magro et al., 21 Nov 2025, Marchant et al., 2022).
- Moment tensor potentials (MTP): Build a hierarchy of scalar invariants from tensor contractions of moment tensors up to high order (Mishin, 2021, Rosenbrock et al., 2019).
- Atomic Cluster Expansion (ACE): Systematic, permutationally and rotationally invariant expansion based on radial polynomials and spherical harmonics, applicable up to arbitrary body order and chemical complexity (Mishin, 2021, Chen et al., 2021).
Regression Models:
- Kernel methods (GAP, Kriging): Implemented as kernel ridge regression or Gaussian process regression, with weights determined by minimizing squared error to a reference quantum dataset (Mishin, 2021, Robredo-Magro et al., 21 Nov 2025, Fellman et al., 2024).
- Feed-forward and message-passing neural networks: Used in BPNN (Rili, 2024), high-dimensional neural network potentials (HDNNP), and modern equivariant GNN architectures such as NequIP, Allegro, MACE, and HIPNN (Bigi et al., 22 Jan 2026, Leimeroth et al., 5 May 2025, Brunken et al., 28 May 2025, Matin et al., 18 Mar 2025, Robredo-Magro et al., 21 Nov 2025).
- Linear polynomial models (SNAP, MTP): Linear regression on polynomial or bispectrum descriptors, with explicit control over model complexity (Mishin, 2021, Rosenbrock et al., 2019).
2. Training Protocols, Data Generation, and Loss Functions
Data Selection and Sampling:
- Accurate and transferable MLIPs require representative, diverse training sets covering relevant regions of configuration space. Automated approaches include information-entropy maximization, leverage-score subsampling, and active-learning protocols that couple MD exploration with uncertainty quantification to supplement training data only where the current MLIP is most uncertain (Baghishov et al., 6 Jun 2025, Kang et al., 2024, Verdi et al., 2021, Allen et al., 2022).
Labeling:
- Energies and forces (and, in some applications, stresses/virials) are computed from ab initio methods (DFT, CCSD(T), etc.) at variable levels of electronic structure precision (Baghishov et al., 6 Jun 2025, Matin et al., 18 Mar 2025). The choice of electronic structure ("reference level") sets a floor on attainable MLIP accuracy.
Loss Functions and Regularization:
- Training most often minimizes a composite loss combining energy and force errors: with user-controlled weighting to balance energy/force accuracy. Recent developments incorporate physically-informed losses enforcing Taylor-consistency between energies and forces as well as energy conservation or symmetry-restoration terms (Takamoto et al., 2024, Matin et al., 18 Mar 2025, Bigi et al., 22 Jan 2026).
Model Complexity:
- Linear and nonlinear models are parameterized by the number of body orders, basis functions, or hidden channels. The choice is guided by the desired trade-off among accuracy, evaluation speed, and required training set size (Baghishov et al., 6 Jun 2025, Leimeroth et al., 5 May 2025, Byggmästar et al., 2022).
3. Model Evaluation, Computational Efficiency, and Software
Quantitative Benchmarks:
- MLIPs achieve typical interpolation errors of 1–10 meV/atom in energy and 20–100 meV/Å in forces for solids and molecules, matching or surpassing classical force fields by $1$–$2$ orders of magnitude and approaching quantum reference accuracy (Leimeroth et al., 5 May 2025, Robredo-Magro et al., 21 Nov 2025, Byggmästar et al., 2022, Fellman et al., 2024).
- Systematic benchmarking across paradigms (GAP, ACE, MTP, NequIP, Allegro, MACE, HDNNP, BPNN) on test problems reveals that nonlinear ACE, equivariant GNNs (NequIP, MACE, Allegro), and tabulated low-dimensional GAP models form the Pareto front in accuracy vs. computational cost (Leimeroth et al., 5 May 2025, Fellman et al., 2024, Byggmästar et al., 2022).
Evaluation Speed and Tabulation:
- High-dimensional kernel models (e.g., SOAP-GAP) are accurate but computationally intensive. Tabulation of low-dimensional MLIP terms (e.g., 2-body, 3-body, EAM-like) with spline-interpolated lookup prior to MD reduces evaluation cost by $1-2$ orders of magnitude without loss of accuracy (Marchant et al., 2022, Fellman et al., 2024, Byggmästar et al., 2022, Byggmästar et al., 2022).
- GPU acceleration of GNN-based MLIPs enables sub-microsecond-per-atom-per-step simulations, rendering million-atom, long-time MD feasible (Brunken et al., 28 May 2025, Leimeroth et al., 5 May 2025).
Integration and Software:
- Standardized software frameworks and libraries now exist for development, training, and deployment of MLIPs (e.g., QUIP/ASE, mlip, open-source LAMMPS/ML-MIX integration) (Brunken et al., 28 May 2025, Birks et al., 26 Feb 2025). These expose APIs for various model types and molecular dynamics engines, including JAX-MD and ASE.
4. Advanced Methods: Multiscale Coupling, Hybridization, and Active Learning
Spatial Mixing and Multiscale Schemes:
- The ML-MIX approach enables spatially inhomogeneous application of MLIPs by mixing expensive (high-accuracy) and cheap (low-order/fitted for efficiency) MLIPs over the simulation domain, via smooth per-atom weighting functions. This reduces the computational cost in large-scale defect or catalysis simulations by up to 11 for 8000 atoms and even greater in larger domains, without significant loss of accuracy in the "active" regions (Birks et al., 26 Feb 2025).
- Constrained linear fitting of the cheap MLIP (enforcing exact elastic constants and optimal reproduction of the expensive reference in bulk configurations) ensures thermodynamic and mechanical consistency within the blended region (Birks et al., 26 Feb 2025).
- Hybrid QM/MM–MLIP coupling for crystalline defects can be rigorously error-analyzed, with controllable error bounds via Taylor and virial matching in the MM/MLIP region, using atomic cluster expansion descriptors and linear least-squares fitting (Chen et al., 2021).
Physically Informed and Weakly Supervised Learning:
- Physics-informed losses (Taylor expansion, conservative-force consistency) injected into the training objective eliminate unphysical force/energy artifacts and yield improved MD stability, even when training labels are sparse or forces are missing (Takamoto et al., 2024).
- Ensemble knowledge distillation allows energies-only datasets (e.g., high-level quantum chemistry where calculating forces is infeasible) to provide force supervision to student MLIPs via ensemble-averaged teacher models, enhancing both energy and force accuracy (Matin et al., 18 Mar 2025).
Transferability and Minimalist Approaches:
- Studies challenge the prevailing notion that exhaustive datasets and hyperparameter optimization are necessary; minimalist MLIPs trained on limited, auto-selected datasets can discover nontrivial structures (e.g., new polymorphs, topological textures) well outside the training domain, especially when descriptors are sufficiently expressive (e.g., SOAP, Allegro) (Robredo-Magro et al., 21 Nov 2025).
Active Learning for Strongly Anharmonic Regimes:
- Active learning loops and ensemble uncertainty estimates, when combined with MLIP-MD, enable robust coverage of strongly anharmonic configuration space and prevent failures such as missing/fake metastable states—crucial for accurate thermal transport and rare-event kinetics (Kang et al., 2024, Verdi et al., 2021).
5. Applications and Demonstrated Impact
Static and Dynamic Properties:
- MLIPs are now reliably used for property prediction (phonons, elastic constants, defects, surfaces), phase diagram determination, phase transitions, and high-fidelity MD in complex systems (e.g., finite-T polymorphism in oxides and metals, defect migration in semiconductors, and carbon phase diagrams over 1 TPa) (Verdi et al., 2021, Marchant et al., 2022, Fellman et al., 2024, Robredo-Magro et al., 21 Nov 2025, Rosenbrock et al., 2019).
Scale Bridging:
- Through tabulation, spatial mixing, and robust GPU-parallel integration, MLIPs enable million-atom simulations over nanosecond timescales, making them competitive or superior in speed to some classical force fields, but with chemical accuracy (Birks et al., 26 Feb 2025, Leimeroth et al., 5 May 2025, Fellman et al., 2024).
Performance Table: Model Cost and Accuracy (example summary)
| Model type | Energy RMSE (meV/atom) | Force RMSE (meV/Å) | Eval Time (μs/atom·step) | Reference |
|---|---|---|---|---|
| Nonlinear ACE | 8–12 | 30–35 | 0.2–0.4 | (Leimeroth et al., 5 May 2025) |
| NequIP/MACE/Allegro GNNs | 6–14 | 25–55 | 4.5–6 | (Leimeroth et al., 5 May 2025Robredo-Magro et al., 21 Nov 2025) |
| Tabulated low-dim GAP | ~2–3 | ~40–100 | 0.1–0.3 | (Byggmästar et al., 2022Fellman et al., 2024) |
| SOAP-GAP | 3–8 | 25–80 | 2.0–4.0 | (Byggmästar et al., 2022Robredo-Magro et al., 21 Nov 2025) |
| Classical EAM/MEAM | 15–100 | 200–600 | 0.005–0.1 | (Leimeroth et al., 5 May 2025Marchant et al., 2022) |
Values are material and system dependent. Full tables with details in (Leimeroth et al., 5 May 2025, Byggmästar et al., 2022, Fellman et al., 2024, Rosenbrock et al., 2019).
6. Practical Guidelines and Recommendations
- For material-specific applications, select MLIP complexity and training data precision to match the accuracy–cost target. Employ leverage sampling or active learning to minimize unnecessary DFT labeling (Baghishov et al., 6 Jun 2025, Kang et al., 2024).
- Use tabulated/low-order descriptors for large-scale, high-throughput, or multicomponent alloy MD; reserve large descriptor sets (e.g., SOAP, GNNs) for cases requiring the highest accuracy or broad chemical transferability (Byggmästar et al., 2022, Leimeroth et al., 5 May 2025).
- Enforce physical constraints (e.g., conservation, elastic properties) when necessary via constrained fitting, physics-informed loss, or post hoc symmetry-restoring corrections (Takamoto et al., 2024, Bigi et al., 22 Jan 2026, Chen et al., 2021, Birks et al., 26 Feb 2025).
- Integrate spatial mixing (ML/ML or QM/ML/MM), as in ML-MIX, when only subregions require full accuracy, enabling domain sizes and timescales far beyond purely homogeneous MLIP-MD (Birks et al., 26 Feb 2025).
- For active discovery, minimalist training and on-the-fly Bayesian or error-driven data expansion can yield MLIPs suitable as "discovery engines" for new polymorphs or cooperative phases (Robredo-Magro et al., 21 Nov 2025, Verdi et al., 2021).
Machine-learned interatomic potentials now form the backbone of high-accuracy, large-scale atomistic simulation, bridging first-principles and empirical methods with theory-driven design, scalable architectures, and integrated practical workflows (Leimeroth et al., 5 May 2025, Mishin, 2021, Birks et al., 26 Feb 2025, Robredo-Magro et al., 21 Nov 2025).