Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Machine-learning interatomic potentials from a users perspective: A comparison of accuracy, speed and data efficiency (2505.02503v1)

Published 5 May 2025 in cond-mat.mtrl-sci

Abstract: Machine learning interatomic potentials (MLIPs) have massively changed the field of atomistic modeling. They enable the accuracy of density functional theory in large-scale simulations while being nearly as fast as classical interatomic potentials. Over the last few years, a wide range of different types of MLIPs have been developed, but it is often difficult to judge which approach is the best for a given problem setting. For the case of structurally and chemically complex solids, namely Al-Cu-Zr and Si-O, we benchmark a range of machine learning interatomic potential approaches, in particular, the Gaussian approximation potential (GAP), high-dimensional neural network potentials (HDNNP), moment tensor potentials (MTP), the atomic cluster expansion (ACE) in its linear and nonlinear version, neural equivariant interatomic potentials (NequIP), Allegro, and MACE. We find that nonlinear ACE and the equivariant, message-passing graph neural networks NequIP and MACE form the Pareto front in the accuracy vs. computational cost trade-off. In case of the Al-Cu-Zr system we find that MACE and Allegro offer the highest accuracy, while NequIP outperforms them for Si-O. Furthermore, GPUs can massively accelerate the MLIPs, bringing them on par with and even ahead of non-accelerated classical interatomic potentials (IPs) with regards to accessible timescales. Finally, we explore the extrapolation behavior of the corresponding potentials, probe the smoothness of the potential energy surfaces, and finally estimate the user friendliness of the corresponding fitting codes and molecular dynamics interfaces.

Summary

We haven't generated a summary for this paper yet.