Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TeaNet: universal neural network interatomic potential inspired by iterative electronic relaxations (1912.01398v2)

Published 2 Dec 2019 in physics.comp-ph, cond-mat.mtrl-sci, cs.LG, and stat.ML

Abstract: A universal interatomic potential for an arbitrary set of chemical elements is urgently needed in computational materials science. Graph convolution neural network (GCN) has rich expressive power, but previously was mainly employed to transport scalars and vectors, not rank $\ge 2$ tensors. As classic interatomic potentials were inspired by tight-binding electronic relaxation framework, we want to represent this iterative propagation of rank $\ge 2$ tensor information by GCN. Here we propose an architecture called the tensor embedded atom network (TeaNet) where angular interaction is translated into graph convolution through the incorporation of Euclidean tensors, vectors and scalars. By applying the residual network (ResNet) architecture and training with recurrent GCN weights initialization, a much deeper (16 layers) GCN was constructed, whose flow is similar to an iterative electronic relaxation. Our traning dataset is generated by density functional theory calculation of mostly chemically and structurally randomized configurations. We demonstrate that arbitrary structures and reactions involving the first 18 elements on the periodic table (H to Ar) can be realized satisfactorily by TeaNet, including C-H molecular structures, metals, amorphous SiO${}_2$, and water, showing surprisingly good performance (energy mean absolute error 19 meV/atom) and robustness for arbitrary chemistries involving elements from H to Ar.

Citations (64)

Summary

We haven't generated a summary for this paper yet.