Fast, Expressive SE$(n)$ Equivariant Networks through Weight-Sharing in Position-Orientation Space (2310.02970v3)
Abstract: Based on the theory of homogeneous spaces we derive geometrically optimal edge attributes to be used within the flexible message-passing framework. We formalize the notion of weight sharing in convolutional networks as the sharing of message functions over point-pairs that should be treated equally. We define equivalence classes of point-pairs that are identical up to a transformation in the group and derive attributes that uniquely identify these classes. Weight sharing is then obtained by conditioning message functions on these attributes. As an application of the theory, we develop an efficient equivariant group convolutional network for processing 3D point clouds. The theory of homogeneous spaces tells us how to do group convolutions with feature maps over the homogeneous space of positions $\mathbb{R}3$, position and orientations $\mathbb{R}3 {\times} S2$, and the group $SE(3)$ itself. Among these, $\mathbb{R}3 {\times} S2$ is an optimal choice due to the ability to represent directional information, which $\mathbb{R}3$ methods cannot, and it significantly enhances computational efficiency compared to indexing features on the full $SE(3)$ group. We support this claim with state-of-the-art results -- in accuracy and speed -- on five different benchmarks in 2D and 3D, including interatomic potential energy prediction, trajectory forecasting in N-body systems, and generating molecules via equivariant diffusion models.
- Cormorant: Covariant molecular neural networks. Advances in Neural Information Processing Systems, 32, 2019.
- Jimmy Aronsson. Mathematical Foundations of Equivariant Neural Networks. Chalmers University of Technology, PhD Thesis, 2023.
- The design space of e (3)-equivariant atom-centered interatomic potentials. arXiv preprint arXiv:2205.06643, 2022a.
- Mace: Higher order equivariant message passing neural networks for fast and accurate force fields. Advances in Neural Information Processing Systems, 35:11423–11436, 2022b.
- E (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature communications, 13(1):2453, 2022.
- Erik J Bekkers. B-spline cnns on lie groups. In International Conference on Learning Representations, 2019.
- Roto-translation covariant convolutional networks for medical image analysis. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part I, pp. 440–448. Springer, 2018.
- Lukas Biewald. Experiment tracking with weights and biases, 2020. URL https://www.wandb.com/. Software available from wandb.com.
- Geometric and physical quantities improve e (3) equivariant message passing. In International Conference on Learning Representations, 2021.
- A program to build e (n)-equivariant steerable cnns. In International Conference on Learning Representations, 2021.
- François Chollet. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1251–1258, 2017.
- On the role of gradients for machine learning of molecular energies and forces. Machine Learning: Science and Technology, 1(4):045018, 2020.
- Group equivariant convolutional networks. In International conference on machine learning, pp. 2990–2999. PMLR, 2016.
- A general theory of equivariant cnns on homogeneous spaces. Advances in neural information processing systems, 32, 2019.
- Gauge equivariant mesh cnns: Anisotropic convolutions on geometric graphs. In International Conference on Learning Representations, 2020.
- Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
- Recent geometric flows in multi-orientation image processing via a cartan connection. Handbook of Mathematical Models and Algorithms in Computer Vision and Imaging: Mathematical Imaging and Vision, pp. 1–60, 2021.
- Left-invariant parabolic evolutions on se(2)𝑠𝑒2se(2)italic_s italic_e ( 2 ) and contour enhancement via invertible orientation scores part i: Linear left-invariant diffusion equations on se(2)𝑠𝑒2se(2)italic_s italic_e ( 2 ). Quarterly of Applied Mathematics, 68(2):255–292, 2010a.
- Left-invariant parabolic evolutions on se(2)𝑠𝑒2se(2)italic_s italic_e ( 2 ) and contour enhancement via invertible orientation scores part ii: Nonlinear left-invariant diffusions on invertible orientation scores. Quarterly of applied mathematics, 68(2):293–331, 2010b.
- On the universality of rotation equivariant point cloud networks. In International Conference on Learning Representations, 2020.
- Fast graph representation learning with PyTorch Geometric. In ICLR Workshop on Representation Learning on Graphs and Manifolds, 2019.
- Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. In International Conference on Machine Learning, pp. 3165–3176. PMLR, 2020.
- Se (3)-transformers: 3d roto-translation equivariant attention networks. Advances in neural information processing systems, 33:1970–1981, 2020.
- E (n) equivariant normalizing flows. Advances in Neural Information Processing Systems, 34:4181–4192, 2021.
- Directional message passing for molecular graphs. In International Conference on Learning Representations, 2019.
- Gemnet: Universal directional graph neural networks for molecules. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=HS_sOaxS9K-.
- Symmetry-adapted generation of 3d point sets for the targeted discovery of molecules. Advances in neural information processing systems, 32, 2019.
- e3nn: Euclidean neural networks, 2022. URL https://arxiv.org/abs/2207.09453.
- Neural message passing for quantum chemistry. In International conference on machine learning, pp. 1263–1272. PMLR, 2017a.
- Neural message passing for quantum chemistry. In Doina Precup and Yee Whye Teh (eds.), Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pp. 1263–1272. PMLR, 06–11 Aug 2017b. URL https://proceedings.mlr.press/v70/gilmer17a.html.
- Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851, 2020.
- Equivariant diffusion for molecule generation in 3d. In International conference on machine learning, pp. 8867–8887. PMLR, 2022.
- Design and processing of invertible orientation scores of 3d images. Journal of mathematical imaging and vision, 60:1427–1458, 2018.
- Variational diffusion models. Advances in neural information processing systems, 34:21696–21707, 2021.
- Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Neural relational inference for interacting systems. In International conference on machine learning, pp. 2688–2697. PMLR, 2018.
- Exploiting redundancy: Separable group convolutional networks on lie groups. In International Conference on Machine Learning, pp. 11359–11386. PMLR, 2022.
- Equivariant flows: exact likelihood generative learning for symmetric densities. In International conference on machine learning, pp. 5361–5370. PMLR, 2020.
- An exploration of conditioning methods in graph neural networks. In ICLR 2023-Machine Learning for Drug Discovery workshop, 2023.
- On the generalization of equivariance and convolution in neural networks to the action of compact groups. In International Conference on Machine Learning, pp. 2747–2755. PMLR, 2018.
- Clebsch–gordan nets: a fully fourier space spherical convolutional neural network. Advances in Neural Information Processing Systems, 31, 2018.
- Regular se (3) group convolutions for volumetric medical image analysis. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2023. arXiv preprint arXiv:2306.13960, 2023.
- Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
- Spherical message passing for 3d molecular graphs. In International Conference on Learning Representations, 2022a. URL https://openreview.net/forum?id=givsRXsOt9r.
- A convnet for the 2020s. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11976–11986, 2022b.
- Learning local equivariant representations for large-scale atomistic dynamics. Nature Communications, 14(1):579, 2023.
- Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Hyena hierarchy: Towards larger convolutional language models. arXiv preprint arXiv:2302.10866, 2023.
- New approximation of a scale space kernel on se (3) and applications in neuroimaging. In International Conference on Scale Space and Variational Methods in Computer Vision, pp. 40–52. Springer, 2015.
- Informing geometric deep learning with electronic interactions to accelerate quantum chemistry. Proceedings of the National Academy of Sciences, 119(31):e2205221119, 2022.
- Rupp M.and Von Lilienfeld O.A. Ramakrishnan R., Dral P.O. Quantum chemistry structures and properties of 134 kilo molecules, 2014.
- Ckconv: Continuous kernel convolution for sequential data. In International Conference on Learning Representations, 2021.
- Clifford group equivariant neural networks. arXiv preprint arXiv:2305.11141, 2023.
- E (n) equivariant graph neural networks. In International conference on machine learning, pp. 9323–9332. PMLR, 2021.
- Schnetpack 2.0: A neural network toolbox for atomistic machine learning. The Journal of Chemical Physics, 158(14), 2023.
- Symmetry-aware actor-critic for 3d molecular design. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=jEYKjPE1xYN.
- Graphvae: Towards generation of small graphs using variational autoencoders. In Artificial Neural Networks and Machine Learning–ICANN 2018: 27th International Conference on Artificial Neural Networks, Rhodes, Greece, October 4-7, 2018, Proceedings, Part I 27, pp. 412–422. Springer, 2018.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp. 2256–2265. PMLR, 2015.
- Maximum likelihood training of score-based diffusion models. Advances in Neural Information Processing Systems, 34:1415–1428, 2021.
- Scale-equivariant steerable networks. In International Conference on Learning Representations, 2019.
- Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, 33:7537–7547, 2020.
- Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds. arXiv preprint arXiv:1802.08219, 2018.
- Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Midi: Mixed graph and 3d denoising diffusion for molecule generation. arXiv preprint arXiv:2302.09048, 2023.
- Scalars are universal: Equivariant machine learning, structured like classical physics. Advances in Neural Information Processing Systems, 34:28848–28863, 2021.
- Visnet: a scalable and accurate geometric deep learning potential for molecular dynamics simulation. arXiv preprint arXiv:2210.16518, 2022.
- General e (2)-equivariant steerable cnns. Advances in neural information processing systems, 32, 2019.
- Learning steerable filters for rotation equivariant cnns. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 849–858, 2018.
- Coordinate independent convolutional networks–isometry and gauge equivariant convolutions on riemannian manifolds. arXiv preprint arXiv:2106.06020, 2021.
- Equivariant and Coordinate Independent Convolutional Networks. 2023. URL https://maurice-weiler.gitlab.io/cnn_book/EquivariantAndCoordinateIndependentCNNs.pdf.
- Cubenet: Equivariance to 3d rotation and translation. In Proceedings of the European Conference on Computer Vision (ECCV), pp. 567–584, 2018.
- Geodiff: A geometric diffusion model for molecular conformation generation. arXiv preprint arXiv:2203.02923, 2022.