Machine Learning Potentials (MLPs)
- Machine Learning Potentials are data-driven surrogate models that simulate atomic interactions with quantum-mechanical accuracy and statistical scalability.
- They employ information-theoretic motif sampling to uniformly cover diverse chemical environments, enhancing predictive transferability across compositions.
- The PACE framework uses a linear atomic cluster expansion with advanced regularization to achieve efficient training and precise energy and force predictions.
Machine learning potentials (MLPs) are data-driven surrogate models that enable atomistic simulations with quantum-mechanical accuracy and statistical-mechanics scalability. MLPs have revolutionized structure/property prediction, dynamics, and phase diagram determination in an array of chemically complex and disordered materials, including multicomponent alloys. A central challenge in this domain is achieving predictive fidelity across the full compositional and structural landscape of such alloys, from perfectly ordered stoichiometric compounds to maximally disordered solid solutions. The integration of information-theoretic sampling strategies with advanced body-ordered descriptors—most notably, the Performant Atomic Cluster Expansion (PACE)—provides an effective means of constructing robust, transferable MLPs for alloys that span this range, as exemplified by recent developments in motif-based sampling (MBS) (Sheriff et al., 14 Jun 2025).
1. Information-Theoretic Sampling of Local Chemical Motifs
The critical link between the local environment of an atom and the macroscopic properties of a material motivates a formal definition of "motifs": each atom’s first coordination polyhedron, labeled by the species of its nearest neighbors, defines a discrete motif . In a database of local environments, the empirical motif frequency is
The combinatorial space of possible motifs for a multicomponent system is vast, and uniform coverage is neither guaranteed nor trivial to realize with random structure enumeration, especially for rare or complex motifs associated with certain compositions or short-range orders.
To rectify sampling bias, the motif distribution is explicitly compared to the uniform reference (with the total number of motifs) using the Jensen–Shannon divergence,
where and .
Maximizing the Shannon entropy , which is equivalent to minimizing , is operationalized through the motif-based sampling (MBS) algorithm:
- Generate an initial pool of alloy structures across the full composition simplex (e.g., fcc random substitutions).
- Within each structure, perform atom swaps to increase .
- Iterate until is acceptably close to uniform across .
This approach ensures that even rare coordination environments—including those present in ordered or highly short-range-ordered regimes—are systematically incorporated into the training set.
2. Model Architecture: The Linear Atomic Cluster Expansion (PACE)
The PACE framework implements a linear atomic cluster expansion (ACE), which provides a systematically improvable, permutation- and rotation-invariant decomposition of the total energy: where are body-ordered cluster basis functions built from one- through -body descriptors of atom ’s neighbor coordinates . Each incorporates radial basis functions , spherical harmonics , and Chebyshev polynomials. For two-body terms,
Higher-order (three-body and beyond) invariants are formed via tensor contractions of these . The PACE codebase automates construction and pruning of the resulting large basis set, ensuring computational tractability and expressivity.
3. Training Workflow and Regularization
PACE-based MLPs are trained by simultaneous matching of reference DFT energies and forces: with and tuned so that energy and force residuals per atom contribute comparably in magnitude (empirically, , ). Tikhonov (ridge) regularization on the coefficients prevents overfitting; – is typically sufficient.
Key features of the workflow include isotropic expansion to mimic finite temperature (up to 1000 K), random displacements, and small strains to sample vibrational anharmonicity.
With MBS, the final dataset achieves a motif "packing density" (fraction of possible motifs present) up to 30% higher than with random sampling and achieves reductions of up to 0.12 bits (e.g., for CrCoNi with 702 structures).
4. Validation: Accuracy and Transferability across Compositions
Quantitative property prediction benchmarks illustrate the consequences of motif entropy sampling:
- Energy/Force Accuracy (CrCoNi): MBS-trained MLPs achieve meV/atom ( meV/atom variation, even at high SRO), compared with $3$–$5$ meV/atom and meV variation for randomly sampled MLPs; force RMSE is consistently eV/Å.
- Phase Diagrams: Predicted fcc–bcc transition boundaries for Cr–Ni and Cr–Co binaries match experiment and CALPHAD within 25–30 K; Au–Pt miscibility gap critical points deviate by only 5% from early experimental data.
- Melting Points: For fcc CrCoNi, K (3% below experiment). For bcc TaTiVW and derivatives, melting predictions are within 2–5% of experimental windows.
- Short-Range Order and Lattice Expansion: Predicted Warren–Cowley and for various compositions and temperatures are within experimental uncertainty ( in SRO, in lattice parameter).
- Thermodynamics: Specific heats () of TaTiVW alloys are predicted within 2–13% of high-temperature NIST and other measurements.
- Fault Energies: Stacking-fault energies in CrCoNi at 500 K map correctly and match first-principles trends as Cr is reduced.
Crucially, MBS-MLPs reduce composition-dependent variation of energy and force errors by two orders of magnitude compared to broader “universal” MLPs (MatterSim, Orb, MACE).
5. Computational Requirements and Scaling
The cost of motif-based sampling and dataset construction is negligible relative to reference quantum mechanical calculations. For example, a 702-structure MBS dataset can be built in less than 10% of the CPU time needed for a single-composition DFT–MC trajectory, yet delivers substantially improved transferability. PACE MLPs exhibit linear scaling () with system size, and inference is efficient enough to support atomistic molecular dynamics of up to millions of atoms.
Dataset sizes for coverage (examples from the work):
- Au–Pt: configurations across stoichiometries and SRO.
- Cu–Au: configurations in analogous sampling.
- Cr–Co–Ni: 66 (benchmark) and 702 (production) structures covering 12 compositions, multiple phases, and stoichiometric compounds.
- Ti–Ta–V–W: configurations targeting equiatomic and derivative compositions.
6. Implications, Extensions, and Future Perspectives
The direct enforcement of motif uniformity achieves order-of-magnitude improvements in compositional transferability, enabling a single MLP to quantitatively predict properties from binary subsystems to high-entropy phases and even liquids without retraining. The motif-based approach is system-agnostic and direct application to other metallic multicomponent alloys, intermetallics, and even non-metallic systems is feasible, contingent on the definition of an appropriate motif alphabet.
Prospective extensions include integration with generative structure proposals to automate the exploration and inclusion of relevant metastable phases. The methodology lays a foundation for the systematic construction of interatomic potentials that are both data-efficient and physically robust across the full compositional and structural landscape—directly addressing a principal limitation of prior universal or database-trained models.
In summary, the combination of information-theoretic sampling (maximally uniform motif entropy) with robust, body-ordered ACE models (PACE) yields a practical and efficient strategy for constructing transferable, high-fidelity MLPs for alloys. These models simultaneously capture chemical diversity, structural complexity, and thermal perturbations, providing a rigorous foundation for predictive atomistic simulation across phase diagrams, property trends, and order–disorder regimes.