Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enabling Efficient Equivariant Operations in the Fourier Basis via Gaunt Tensor Products (2401.10216v2)

Published 18 Jan 2024 in cs.LG, cond-mat.mtrl-sci, math.GR, physics.chem-ph, and q-bio.BM

Abstract: Developing equivariant neural networks for the E(3) group plays an important role in modeling 3D data across real-world applications. Enforcing this equivariance primarily involves the tensor products of irreducible representations (irreps). However, the computational complexity of such operations increases significantly as higher-order tensors are used. In this work, we propose a systematic approach to substantially accelerate the computation of the tensor products of irreps. We mathematically connect the commonly used Clebsch-Gordan coefficients to the Gaunt coefficients, which are integrals of products of three spherical harmonics. Through Gaunt coefficients, the tensor product of irreps becomes equivalent to the multiplication between spherical functions represented by spherical harmonics. This perspective further allows us to change the basis for the equivariant operations from spherical harmonics to a 2D Fourier basis. Consequently, the multiplication between spherical functions represented by a 2D Fourier basis can be efficiently computed via the convolution theorem and Fast Fourier Transforms. This transformation reduces the complexity of full tensor products of irreps from $\mathcal{O}(L6)$ to $\mathcal{O}(L3)$, where $L$ is the max degree of irreps. Leveraging this approach, we introduce the Gaunt Tensor Product, which serves as a new method to construct efficient equivariant operations across different model architectures. Our experiments on the Open Catalyst Project and 3BPA datasets demonstrate both the increased efficiency and improved performance of our approach.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (87)
  1. Geometric deep learning on molecular representations. Nature Machine Intelligence, 3(12):1023–1032, 2021.
  2. Gaussian approximation potentials: The accuracy of quantum mechanics, without the electrons. Physical review letters, 104(13):136403, 2010.
  3. The design space of e (3)-equivariant atom-centered interatomic potentials. arXiv preprint arXiv:2205.06643, 2022a.
  4. Mace: Higher order equivariant message passing neural networks for fast and accurate force fields. Advances in Neural Information Processing Systems, 35:11423–11436, 2022b.
  5. E (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature communications, 13(1):2453, 2022.
  6. Generalized neural-network representation of high-dimensional potential-energy surfaces. Physical review letters, 98(14):146401, 2007.
  7. Geometric and physical quantities improve e(3) equivariant message passing. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=_xwr8gOBeV1.
  8. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021.
  9. Gerald Burns. Introduction to group theory with applications: materials science and technology. Academic Press, 2014.
  10. Open catalyst 2020 (oc20) dataset and community challenges. ACS Catalysis, 11(10):6059–6072, 2021.
  11. GeoMFormer: A general architecture for geometric molecular representation learning. In NeurIPS 2023 AI for Science Workshop, 2023. URL https://openreview.net/forum?id=s0UNtuuqU5.
  12. Machine learning of accurate energy-conserving molecular force fields. Science advances, 3(5):e1603015, 2017.
  13. sgdml: Constructing accurate and data efficient molecular force fields using machine learning. Computer Physics Communications, 240:38–45, 2019.
  14. Group equivariant convolutional networks. In International conference on machine learning, pp.  2990–2999. PMLR, 2016.
  15. Steerable CNNs. In International Conference on Learning Representations, 2017. URL https://openreview.net/forum?id=rJQKYt5ll.
  16. Eugene D Commins. Modern quantum mechanics, revised edition, 1995.
  17. Spherenet: Learning spherical representations for detection and classification in omnidirectional images. In Proceedings of the European conf. on computer vision (ECCV), 2018.
  18. John F Cornwell. Group theory in physics: An introduction. Academic press, 1997.
  19. F Albert Cotton. Chemical applications of group theory. John Wiley & Sons, 1991.
  20. Vector neurons: A general framework for so (3)-equivariant networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  12200–12209, 2021.
  21. Ralf Drautz. Atomic cluster expansion for accurate and transferable interatomic potentials. Physical Review B, 99(1):014104, 2019.
  22. Atomic cluster expansion: Completeness, efficiency and stability. Journal of Computational Physics, 454:110946, 2022.
  23. A hitchhiker’s guide to geometric gnns for 3d atomic systems. arXiv preprint arXiv:2312.07511, 2023.
  24. On the universality of rotation equivariant point cloud networks. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=6NFBvWlRXaG.
  25. Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. In International Conference on Machine Learning, pp.  3165–3176. PMLR, 2020.
  26. So3krates: Equivariant attention for interactions on arbitrary length-scales in molecular systems. Advances in Neural Information Processing Systems, 35:29400–29413, 2022.
  27. Se (3)-transformers: 3d roto-translation equivariant attention networks. Advances in neural information processing systems, 33:1970–1981, 2020.
  28. Fast and uncertainty-aware directional message passing for non-equilibrium molecules. In NeurIPS-W, 2020a.
  29. Directional message passing for molecular graphs. In International Conference on Learning Representations, 2020b. URL https://openreview.net/forum?id=B1eWbxStPH.
  30. Gemnet: Universal directional graph neural networks for molecules. Advances in Neural Information Processing Systems, 34:6790–6802, 2021.
  31. Gemnet-oc: developing graph neural networks for large and diverse molecular simulation datasets. Transactions on Machine Learning Research, 2022.
  32. e3nn: Euclidean neural networks. arXiv preprint arXiv:2207.09453, 2022.
  33. General framework for e (3)-equivariant neural network representation of density functional theory hamiltonian. Nature Communications, 14(1):2848, 2023.
  34. Newtonnet: A newtonian message passing network for deep learning of interatomic potentials and forces. Digital Discovery, 1(3):333–343, 2022.
  35. Improved adsorption energetics within density-functional theory using revised perdew-burke-ernzerhof functionals. Physical review B, 59(11):7413, 1999.
  36. Geometrically equivariant graph neural networks: A survey. arXiv preprint arXiv:2202.07230, 2022.
  37. Deep networks with stochastic depth. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part IV 14, pp.  646–661. Springer, 2016.
  38. Lietransformer: Equivariant self-attention for lie groups. In International Conference on Machine Learning, pp.  4533–4543. PMLR, 2021.
  39. Sigeru Huzinaga. Molecular integrals. Progress of Theoretical Physics Supplement, 40:52–77, 1967.
  40. Nadir Jeevanjee. An introduction to tensors and group theory for physicists. Springer, 2011.
  41. Learning from protein structure with geometric vector perceptrons. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=1YLJDvSx6J4.
  42. On the expressive power of geometric graph neural networks. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  15330–15355. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/joshi23a.html.
  43. Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589, 2021.
  44. nabladft: Large-scale conformational energy and hamiltonian prediction benchmark and dataset. Physical Chemistry Chemical Physics, 24(42):25853–25863, 2022.
  45. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  46. Yvette Kosmann-Schwarzbach et al. Groups and symmetries. Springer, 2010.
  47. Linear atomic cluster expansion force fields for organic molecules: beyond rmse. Journal of chemical theory and computation, 17(12):7696–7711, 2021.
  48. Evaluation of the MACE force field architecture: From medicinal chemistry to materials science. The Journal of Chemical Physics, 159(4):044118, 07 2023. ISSN 0021-9606. doi: 10.1063/5.0155322. URL https://doi.org/10.1063/5.0155322.
  49. Adsorbml: Accelerating adsorption energy calculations with machine learning. arXiv preprint arXiv:2211.16486, 2022.
  50. Deep-learning density functional theory hamiltonian for efficient ab initio electronic-structure calculation. Nature Computational Science, 2(6):367–377, 2022.
  51. Equiformer: Equivariant graph attention transformer for 3d atomistic graphs. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=KwmPfARgOTD.
  52. Equiformerv2: Improved equivariant transformer for scaling to higher-degree representations. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=mCOBKZmrzD.
  53. One transformer can understand both 2d & 3d molecular data. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=vZTp1oPV3PC.
  54. Learning local equivariant representations for large-scale atomistic dynamics. Nature Communications, 14(1):579, 2023.
  55. Unified theory of atom-centered representations and message-passing machine-learning schemes. The Journal of Chemical Physics, 156(20), 2022.
  56. Reducing SO(3) convolutions to SO(2) for efficient equivariant GNNs. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  27420–27438. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/passaro23a.html.
  57. Richard Pio. Euler angle transformations. IEEE Transactions on automatic control, 11(4):707–715, 1966.
  58. John G Proakis. Digital signal processing: principles, algorithms, and applications, 4/E. Pearson Education India, 2007.
  59. Quantum chemistry structures and properties of 134 kilo molecules. Scientific data, 1(1):1–7, 2014.
  60. Joseph J Rotman. An introduction to the theory of groups. Graduate Texts in Mathematics, 1995.
  61. E (n) equivariant graph neural networks. In International conference on machine learning, pp.  9323–9332. PMLR, 2021.
  62. Equivariant message passing for the prediction of tensorial properties and molecular spectra. In International Conference on Machine Learning, pp.  9377–9388. PMLR, 2021.
  63. Schnet–a deep learning architecture for molecules and materials. The Journal of Chemical Physics, 148(24):241722, 2018.
  64. Alexander V Shapeev. Moment tensor potentials: A class of systematically improvable interatomic potentials. Multiscale Modeling & Simulation, 14(3):1153–1173, 2016.
  65. Benchmarking graphormer on large-scale molecular modeling datasets. arXiv preprint arXiv:2203.04810, 2022.
  66. Rotation invariant graph neural networks using spin convolutions. arXiv preprint arXiv:2106.09575, 2021.
  67. Tensornet: Cartesian tensor representations for efficient learning of molecular potentials. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=BEHlPdBZ2e.
  68. Arnold Sommerfeld. Partial differential equations in physics. Academic press, 1949.
  69. Equivariant transformers for neural network based molecular potentials. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=zNHzqZ9wrRB.
  70. Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds. arXiv preprint arXiv:1802.08219, 2018.
  71. Spectral neighbor analysis method for automated generation of quantum-accurate interatomic potentials. Journal of Computational Physics, 285:316–330, 2015.
  72. Se (3)-equivariant prediction of molecular wavefunctions and electronic densities. Advances in Neural Information Processing Systems, 34:14434–14447, 2021.
  73. Revisiting point cloud classification: A new benchmark dataset and classification model on real-world data. In Proceedings of the IEEE/CVF international conference on computer vision, pp.  1588–1597, 2019.
  74. Quantum theory of angular momentum. World Scientific, 1988.
  75. Spatial attention kinetic networks with e(n)-equivariance. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=3DIpIf3wQMC.
  76. 3d steerable cnns: Learning rotationally equivariant features in volumetric data. Advances in Neural Information Processing Systems, 31, 2018.
  77. EP Wigner. On the matrices which reduce the kronecker products of representations of sr groups, manuscript (1940), published in lc biedenharn and h. van dam (eds.), quantum theory of angular momentum, 87–133, 1965.
  78. Eugene Wigner. Group theory: and its application to the quantum mechanics of atomic spectra, volume 5. Elsevier, 2012.
  79. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  1912–1920, 2015.
  80. Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems, 34:28877–28888, 2021a.
  81. First place solution of kdd cup 2021 & ogb large-scale challenge graph prediction track. arXiv preprint arXiv:2106.08279, 2021b.
  82. QH9: A quantum hamiltonian prediction benchmark for QM9 molecules. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2023a. URL https://openreview.net/forum?id=71uRr9N39A.
  83. Efficient and equivariant graph networks for predicting quantum Hamiltonian. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  40412–40424. PMLR, 23–29 Jul 2023b. URL https://proceedings.mlr.press/v202/yu23i.html.
  84. Rethinking the expressive power of GNNs via graph biconnectivity. In The Eleventh International Conference on Learning Representations, 2023a. URL https://openreview.net/forum?id=r9hNv76KoT3.
  85. Artificial intelligence for science in quantum, atomistic, and continuum systems. arXiv preprint arXiv:2307.08423, 2023b.
  86. Ya A Granovskiıand AS Zhedanov et al. Nature of the symmetry group of the 6j-symbol. Zh. Eksp. Teor. Fiz., 94:49, 1988.
  87. Spherical channels for modeling atomic interactions. In Advances in Neural Information Processing Systems, 2022.
Citations (15)

Summary

  • The paper presents the Gaunt Tensor Product to reduce computational complexity in equivariant operations from cubic to linear.
  • It leverages Fast Fourier Transforms in a 2D Fourier basis to enable rapid tensor product computations for efficient feature interactions.
  • Experimental validation shows enhanced speed and accuracy in equivariant convolutions and many-body interaction models across diverse architectures.

Understanding Efficiency in Equivariant Neural Networks with the Gaunt Tensor Product

Introduction to Equivariant Neural Networks

Equivariant neural networks are a key advancement in machine learning-related 3D data modeling. These networks are designed to consistently transform their outputs in response to changes in their inputs, maintaining symmetries that naturally occur in data such as molecular structures or 3D images. A typical way to implement equivariance in neural networks involves using tensor products of irreducible representations (irreps), employing tools from representation theory. While powerful, this approach faces computational challenges as operations become complex, especially when dealing with higher-order tensors.

Advancing Computational Efficiency with the Gaunt Tensor Product

A team of researchers has proposed an innovative solution to the computational obstacles posed by conventional methods when tensor products of irreps are used for enforcing equivariance. They introduce the Gaunt Tensor Product, aiming to provide a significant computational speed-up by exploiting the mathematical relationship between the widely used Clebsch-Gordan coefficients and the Gaunt coefficients.

The key insight of this work is that the aforementioned coefficients, central to equivariant operations, can be connected to efficiently computable integrals involving spherical harmonics. By transitioning to a 2D Fourier basis representation, the challenging tensor products can then be rapidly computed through efficient Fast Fourier Transforms (FFTs), leveraging the convolution theorem.

This novel method substantially reduces computational complexity from a curve with a steep incline (cubic in the max degree of irreps) to a more tractable linear growth curve, thereby facilitating the construction of more efficient equivariant operations across different model architectures without sacrificing their capabilities.

Application Across Various Model Architectures

The Gaunt Tensor Product is not only faster but also versatile. The authors delve into three primary classes of operations where their approach can be applied:

  • Equivariant Feature Interactions: It enables features from different dimensional spaces and objects to interact, a fundamental aspect in many advanced models.
  • Equivariant Convolutions: By incorporating insights from advanced models and adding sparsity, their method achieves even greater accelerations for key building blocks of equivariant message passing.
  • Equivariant Many-Body Interactions: Essential for capturing complex interaction effects in molecular systems, their method utilizes a divide-and-conquer strategy to efficiently parallelize computations, making it especially effective for models predicting dynamic systems like force fields.

Experimental Validation and Impacts

Extensive experiments confirm the effectiveness and efficiency of the Gaunt Tensor Product. The approach demonstrates significant speed improvements compared to existing libraries and methods on standard benchmark datasets while maintaining or exceeding modeling accuracy. Through these experiments, the team establishes the Gaunt Tensor Product as a formidable tool for the future of equivariant neural networks, especially in fields requiring large-scale 3D molecular modeling.

The innovative approach promises to ease the computational demands imposed on researchers and pave the way for new exploration opportunities in developing advanced equivariant models. As the field of AI continues to grow, such improvements in computational methods are crucial to harness the full potential of machine learning for understanding and predicting complex physical phenomena.