Charge-Equilibrated TensorNet (QET)
- Charge-Equilibrated TensorNet (QET) is a neural network-based atomistic potential that combines equivariant tensor representations with an analytic charge equilibration scheme.
- The model employs a tensor-based message-passing framework with strict E(3)-equivariance to efficiently compute energies, forces, and reactive electrostatics over large material systems.
- QET achieves linear scaling by replacing cubic charge equilibration with local neighbor summations, enabling accurate simulations in energy storage, catalysis, and interfacial studies.
Charge-Equilibrated TensorNet (QET) is a neural network-based atomistic potential that achieves charge-awareness and strict equivariance with linear scaling, enabling efficient, self-consistent modeling of electrostatics and charge transfer in materials far beyond the computational reach of conventional approaches. QET solves the long-standing challenge of incorporating charge equilibration and reactive electrostatics into foundation potentials for large-scale simulations, allowing for accurate predictions in systems where electron redistribution and voltage effects are critical. The approach combines a tensor-based message-passing network with an analytic, environment-dependent charge equilibration scheme, facilitating applications in energy storage, heterogeneous catalysis, ionic liquids, and complex solid–liquid interfaces (Ko et al., 10 Nov 2025).
1. TensorNet Backbone and Feature Construction
QET extends the TensorNet equivariant graph neural network architecture, in which each atom is endowed with a rank-2 tensor feature . This representation jointly encodes scalar, vector, and higher-order geometric (e.g., quadrupolar) descriptors. The essential stages in a forward pass are:
- Input: Atomic numbers , coordinates , and total system charge .
- Embedding: Each atom receives a tensor feature via a learned embedding.
- Message Passing: Three interaction (update) blocks perform equivariant convolutional message passing over neighbor atoms within a fixed cutoff (5.0 Å), updating for .
- Readout: Each atomic tensor is reduced to a scalar feature vector through contraction and pooling.
- Charge Equilibration (LQeq Block): A multi-layer perceptron (MLP) outputs environment-dependent electronegativity and hardness for each atom, which are passed to the analytic charge solver.
- Augmentation and Energy Head: The final atomic feature (where is the partial charge and a short-range Coulomb sum) is processed by a gated MLP to yield per-atom energy . The total potential energy is .
The architecture preserves exact -equivariance, permitting seamless modeling of bulk, molecular, and interfacial systems under arbitrary transformations.
2. Analytic and Linear-Scaling Charge Equilibration
Central to QET is its charge equilibration scheme, a modification of the classical Qeq formalism with analytic solutions that eliminate the cubic scaling bottleneck. The total energy for a set of partial atomic charges is
subject to .
Here, represents the (short-range) screened Coulomb interaction:
with and the interatomic separation.
Minimizing with a Lagrange multiplier leads to the linear system
where diagonal elements and off-diagonal elements .
QET circumvents the standard matrix inversion ( cost) by introducing an effective electronegativity:
so that
The charge sum constraint gives :
Thus, each is updated in closed form using only local neighbor summations and two global reductions, all scaling as .
3. Training Regimen and Implementation
QET was trained on "MatQ," a dataset of approximately 114,000 structures spanning 86 elements, derived from DFT-PBE calculations with DDEC6 atomic charges. The data encompasses both near-equilibrium (strained crystals from the Materials Project) and out-of-equilibrium (AIMD and rattling) configurations. The loss is a weighted Huber function over energies, forces, stress tensors, and atomic charges. Typical weights for the foundational model were , , , , with different regimens for specialized and fine-tuned tasks.
Regularization includes a softplus transform on to maintain positive hardness and subtraction of elemental reference energies for enhanced convergence. Optimization is performed using AdamW (AMSGrad) with a cosine-annealed learning rate schedule. Early stopping guards against overfitting. The reference implementation is provided via PyTorch Lightning in the Materials Graph Library.
4. Benchmark Performance and Comparative Analysis
QET achieves comparable or superior accuracy to existing charge-aware and charge-agnostic models on standard benchmarks while maintaining linear inference cost. On small charged clusters (e.g., , , , –), QET exhibits mean absolute errors (MAE) that are best-in-class for energies (e.g., 0.64 meV/atom for ), with charge prediction errors matching or surpassing conventional Qeq models.
For bulk crystals, QET-MatQ performs on par with foundation potentials trained on datasets fourfold or more larger. Structure metrics, bulk and shear moduli, and heat capacity estimates are comparable to state-of-the-art, charge-agnostic TensorNet and advanced alternatives.
| System | Energy MAE (meV/atom) | Forces MAE (meV/Å) | Charge MAE (milli-e) |
|---|---|---|---|
| / | 0.64 / 28.5 | 6.58 / 4.92 | 5.66 |
| 0.79 | 14.0 | 1.81 | |
| 0.58 | 27.3 | 11.5 | |
| – | 0.51 | 4.58 | 15.5 |
A key distinction is that QET achieves this performance at a computational cost scaling strictly as , whereas conventional Qeq-based models exhibit cubic or near-quadratic scaling, limiting them to much smaller systems.
5. Unique Capabilities: Charge Transfer and Electrochemical Reactivity
QET's charge-aware architecture allows it to capture physical behaviors inaccessible to charge-agnostic models. In extensive MD simulations of – ionic liquid (1,040 atoms, 100 ps, 1,200 K, NPT), only QET reproduced key experimental observables:
- Structure factors : QET accurately matched experimental peak positions and avoided spurious ordering at low , in contrast to TensorNet and GRACE potentials.
- Density: QET predicted the correct value within 1%, where TensorNet severely overestimated and GRACE underestimated.
These qualitative gains stem from QET’s ability to model non-local, self-consistent electrostatics driving liquid disorder.
For reactive solid–electrolyte interfaces (/), QET detects and maintains the correct phosphorus oxidation states and prevents unphysical bond formation (e.g., – bonds). Under applied bias (±4 V), QET supports MD with dynamic, voltage-responsive charge distributions. Negative bias promotes reduction and continued interphase growth; positive bias hallmarks passivation and structural stasis, reflected in the evolving energy landscape and charge histograms over nanosecond-scale trajectories.
6. Limitations and Prospective Extensions
QET omits long-range Coulomb tail contributions beyond a fixed neighbor cutoff; explicit Ewald or particle-mesh summations can be reintroduced where necessary, at increased (but still sub-cubic) cost. The present model only treats monopolar charge transfer; learning multipole moments and environment-dependent charge widths () would further enhance fidelity. All training references arise from a single DFT functional (PBE) and population scheme (DDEC6); the accuracy of QET is inherently limited by the underlying reference data.
Potential directions include incorporating learned Ewald terms, higher multipoles, or environmental charge distributions in an end-to-end differentiable, equivariant manner. Application-specific fine-tuning can relax the charge loss for situations where no reference partitioning is available.
7. Broader Implications: Quantum Many-Body, Categorical, and Charge Network Generalizations
In the broader context of generalized tensor networks, the original proposal for a "Charge-Equilibrated TensorNet" (QET) also arises as an open challenge in the charged-string (JLW) network formalism (Biamonte, 2017). There, a QET demands dynamic insertion/removal of charge pairs for contraction equilibration, movement beyond abelian fusion (parafermionic qudits), and local charge-conserving rewrite rules to make network contraction efficient and physically meaningful. Key technical barriers in both the quantum information and materials modeling domains include combinatorial proliferation of charge labels, managing phase factors from nontrivial braiding/isotopies, and the handling of non-Abelian or parafermionic structures.
In both materials science and tensor network quantum models, QETs represent a strategy to enforce and exploit local/global charge conservation in scalable simulations. The practical realization in atomistic ML potentials marries analytic, strictly local charge solutions to equivariant message passing, removing the computational bottlenecks that, until recently, limited predictive electrostatics, reactivity, and voltage control to small systems.
Charge-Equilibrated TensorNet establishes a scalable architecture for atomistic simulation, addressing the critical electrostatics bottleneck with applications in energy storage, catalysis, and fundamental studies of charge transfer. Its generalization to other domains and future enhancement via non-abelian extensions or end-to-end learned multipoles remains an active research frontier.