Atomic Foundation Models Overview
- Atomic Foundation Models are large-scale, pre-trained machine learning models that provide physically consistent predictions for energy, force, and properties in atomistic simulations.
- They leverage advanced equivariant architectures and rigorous scaling laws to ensure size-consistent and symmetry-preserving computations across diverse chemical spaces.
- These models enable practical applications in materials science, chemistry, and biology through fine-tuning, transfer learning, and efficient surrogate distillation techniques.
Atomic Foundation Models are large-scale, pre-trained machine learning models designed to deliver general-purpose, physically consistent predictive capabilities for atomistic simulations across materials, molecules, and chemical systems. These models are trained on broad chemical spaces covering many elements and structures, and they provide a unified architecture for accurate energy, force, and property prediction. Atomic Foundation Models inherit principles of scale—expressed through data and parameter scaling laws—robust 3D representations, and advanced equivariant architectures, parallel to the role of LLMs and vision FMs in their respective domains (Yuan et al., 13 Mar 2025, Batatia et al., 2023).
1. Mathematical Formalism and Model Definition
Atomic Foundation Models (AFMs) universally implement an atom-centered energy decomposition: where denotes the atomic configuration, is system size, is the neighbor list for atom (within cutoff ), and are parameters. Forces follow by analytic differentiation: The mapping must exhibit correct invariances—translational, permutational, often rotational equivariance: for any translation and rotation matrix (Yuan et al., 13 Mar 2025, Chen et al., 2024, Li et al., 5 Dec 2025).
2. Scaling Laws, Data Regimes, and Architectural Families
AFMs are governed by neural scaling relations: with exponents , for atomistic problems (Yuan et al., 13 Mar 2025, Wood et al., 30 Jun 2025). UMA models demonstrate compute-optimal scaling for atomic tasks, where model size and dataset size should be balanced according to iso-FLOPs scaling laws (Wood et al., 30 Jun 2025): MoLE layers allow model capacity to increase without slowing inference, enabling high-throughput simulation with billions of parameters but only tens of millions of active parameters per atomic system.
Architectural families include:
- Equivariant graph neural networks (EGNNs): NequIP, Allegro, MACE; tensor interplay enforces E(3) symmetry (Nomura et al., 9 Feb 2025, Yuan et al., 13 Mar 2025).
- Transformer-based models: Attention across atomic embeddings, with some models adopting spherical harmonics–attended blocks.
- Descriptor-based models: ACE, SOAP, bag-of-descriptors, which use symmetry-adapted invariants.
These architectures guarantee strict size-consistency, locality, and physical invariance, and are benchmarked for cross-domain performance (Batatia et al., 2023, Nomura et al., 9 Feb 2025, Wood et al., 30 Jun 2025).
3. Pre-training Strategies and Representation Manifolds
Pre-training of AFMs exploits massive datasets—Materials Project, OMat24, OC20++, OMol25, etc.—with objectives:
- Supervised energy–force regression:
- Self-supervised graph denoising and masked node/edge prediction (Nomura et al., 9 Feb 2025).
To address the lack of interoperability across model architectures, the Platonic representation projects all model-specific atomic embeddings onto a common anchor-defined manifold, using cosine similarities to a DIRECT-sampled anchor basis: yielding a latent space that preserves periodicity, symmetry, and enables model-to-model optimal transport and arithmetic (Li et al., 5 Dec 2025). Embedding arithmetic generalizes to material-level and reaction-level algebra, with compatible cross-model operations.
4. Fine-tuning, Transfer Learning, and Surrogate Distillation
AFMs are adapted to downstream tasks using minimal data and layers modulation:
- Frozen transfer learning: Selectively re-train upper layers (e.g., latest interaction blocks or readout head), keeping most architecture frozen for efficiency (Radova et al., 21 Feb 2025).
- Distillation: Teacher FMs generate synthetic data via rattle-relax, students (e.g., ACE, PaiNN) are fit to FM outputs, achieving up to inference speed-ups with marginal accuracy loss (Gardner et al., 12 Jun 2025).
- Δ-learning via GPR: Residual corrections are learned atop internal model embeddings with Gaussian-process regression, using species- or atomic-level aggregation to fix coverage-induced errors (e.g., underrepresented metal–sulfur chemistry) (Christiansen et al., 28 Feb 2025).
Typical simulation workflows proceed: pre-trained FM → targeted fine-tuning (partial freezing) → validation → surrogate model fit → high-throughput production MD (Radova et al., 21 Feb 2025, Gardner et al., 12 Jun 2025).
5. Model Evaluation, Interoperability, and Bias Detection
Universal metrics are required to assess AFM performance:
- Energy/force/stress RMSEs on held-out sets and OOD splits (Wood et al., 30 Jun 2025, Zheng et al., 17 Jan 2025).
- Procrustes scores, mutual k-NN, optimal transport, SuperScores to quantify representational alignment across architectures (Li et al., 5 Dec 2025).
- Physical error amplification: Hessian-derived phonon errors reveal force-prediction sensitivity, especially for high-curvature PES sectors (Yang et al., 21 Oct 2025).
- Equivariance detection: Fraction of collapsed symmetry-site embeddings (e.g., test) (Li et al., 5 Dec 2025).
The Platonic framework demonstrates how local architectures yield consistent, interpretable global statistics when projected, and allows early diagnostic detection of training biases or symmetry breaking.
6. Domain Coverage, Applications, and Emergent Capabilities
Atomic Foundation Models are validated across:
- Materials properties: formation energies, elastic moduli, phonons, band structures (Batatia et al., 2023, Wood et al., 30 Jun 2025, Yang et al., 21 Oct 2025).
- Chemistry: molecular reactions, thermochemistry, solvation, olfaction, isotope prediction (Wadell et al., 20 Oct 2025).
- Biomolecules: local-environment embeddings from AFMs enable NMR chemical shift prediction, pKa regression, and functional priors over protein environments (Bojan et al., 29 May 2025).
- Inverse design: conditional crystal generation models are integrated with AFMs for property-driven active learning (Li et al., 24 Feb 2025).
- Exascale simulation: strictly-local architectures (e.g., Allegro-FM) achieve billion-atom, nanosecond-scale MD with near-perfect parallel efficiency (Nomura et al., 9 Feb 2025).
Emergent behaviors observed include robust cross-domain generalization, zero/few-shot adaptation, mechanistic interpretability, and algebraic compatibility of embeddings (Li et al., 5 Dec 2025, Wadell et al., 20 Oct 2025, Batatia et al., 2023).
7. Open Challenges and Future Directions
Critical frontiers for Atomic Foundation Models:
- Long-range interactions: Explicit incorporation of electrostatics, dispersion, and dielectric responses (Yuan et al., 13 Mar 2025, Wood et al., 30 Jun 2025).
- Excited states, spin, and charge transfer: Extending architectures for radicals, nonadiabatic dynamics, and charge fluctuations (Yuan et al., 13 Mar 2025).
- Multimodal fusion: Integrating spectroscopic, textual, and microstructural modalities (Van et al., 25 Jun 2025, Wadell et al., 20 Oct 2025).
- Governance, benchmarks, reproducibility: Standardized evaluation (LLM4Mat-Bench, MACBENCH), model zoos, open-data infrastructure (Van et al., 25 Jun 2025, Wood et al., 30 Jun 2025).
- Active & continual learning: Autonomous retraining from closed-loop laboratory data, uncertainty-aware adaptation (Van et al., 25 Jun 2025, Li et al., 24 Feb 2025).
Realizing the full promise of AFMs will require coordinated, large-scale data acquisition, algorithmic innovation for physics-aware learning, scalable compute, and comprehensive benchmarking that measures both foundational capacity and practical simulation utility.
Atomic Foundation Models unify disparate atomistic representations, leveraging massive datasets and equivariant architectures to enable scalable, interpretable, and robust simulations for materials, chemistry, and biological systems. Projection-based frameworks such as the Platonic representation support true interoperability, diagnostic error analysis, and embedding arithmetic across models, positioning AFMs as the scientific infrastructure for next-generation atomistic simulation (Li et al., 5 Dec 2025, Yuan et al., 13 Mar 2025, Wood et al., 30 Jun 2025, Batatia et al., 2023).