Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
140 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fast, Expressive SE$(n)$ Equivariant Networks through Weight-Sharing in Position-Orientation Space (2310.02970v3)

Published 4 Oct 2023 in cs.LG and math.GR

Abstract: Based on the theory of homogeneous spaces we derive geometrically optimal edge attributes to be used within the flexible message-passing framework. We formalize the notion of weight sharing in convolutional networks as the sharing of message functions over point-pairs that should be treated equally. We define equivalence classes of point-pairs that are identical up to a transformation in the group and derive attributes that uniquely identify these classes. Weight sharing is then obtained by conditioning message functions on these attributes. As an application of the theory, we develop an efficient equivariant group convolutional network for processing 3D point clouds. The theory of homogeneous spaces tells us how to do group convolutions with feature maps over the homogeneous space of positions $\mathbb{R}3$, position and orientations $\mathbb{R}3 {\times} S2$, and the group $SE(3)$ itself. Among these, $\mathbb{R}3 {\times} S2$ is an optimal choice due to the ability to represent directional information, which $\mathbb{R}3$ methods cannot, and it significantly enhances computational efficiency compared to indexing features on the full $SE(3)$ group. We support this claim with state-of-the-art results -- in accuracy and speed -- on five different benchmarks in 2D and 3D, including interatomic potential energy prediction, trajectory forecasting in N-body systems, and generating molecules via equivariant diffusion models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (72)
  1. Cormorant: Covariant molecular neural networks. Advances in Neural Information Processing Systems, 32, 2019.
  2. Jimmy Aronsson. Mathematical Foundations of Equivariant Neural Networks. Chalmers University of Technology, PhD Thesis, 2023.
  3. The design space of e (3)-equivariant atom-centered interatomic potentials. arXiv preprint arXiv:2205.06643, 2022a.
  4. Mace: Higher order equivariant message passing neural networks for fast and accurate force fields. Advances in Neural Information Processing Systems, 35:11423–11436, 2022b.
  5. E (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature communications, 13(1):2453, 2022.
  6. Erik J Bekkers. B-spline cnns on lie groups. In International Conference on Learning Representations, 2019.
  7. Roto-translation covariant convolutional networks for medical image analysis. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part I, pp.  440–448. Springer, 2018.
  8. Lukas Biewald. Experiment tracking with weights and biases, 2020. URL https://www.wandb.com/. Software available from wandb.com.
  9. Geometric and physical quantities improve e (3) equivariant message passing. In International Conference on Learning Representations, 2021.
  10. A program to build e (n)-equivariant steerable cnns. In International Conference on Learning Representations, 2021.
  11. François Chollet. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  1251–1258, 2017.
  12. On the role of gradients for machine learning of molecular energies and forces. Machine Learning: Science and Technology, 1(4):045018, 2020.
  13. Group equivariant convolutional networks. In International conference on machine learning, pp.  2990–2999. PMLR, 2016.
  14. A general theory of equivariant cnns on homogeneous spaces. Advances in neural information processing systems, 32, 2019.
  15. Gauge equivariant mesh cnns: Anisotropic convolutions on geometric graphs. In International Conference on Learning Representations, 2020.
  16. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
  17. Recent geometric flows in multi-orientation image processing via a cartan connection. Handbook of Mathematical Models and Algorithms in Computer Vision and Imaging: Mathematical Imaging and Vision, pp.  1–60, 2021.
  18. Left-invariant parabolic evolutions on s⁢e⁢(2)𝑠𝑒2se(2)italic_s italic_e ( 2 ) and contour enhancement via invertible orientation scores part i: Linear left-invariant diffusion equations on s⁢e⁢(2)𝑠𝑒2se(2)italic_s italic_e ( 2 ). Quarterly of Applied Mathematics, 68(2):255–292, 2010a.
  19. Left-invariant parabolic evolutions on s⁢e⁢(2)𝑠𝑒2se(2)italic_s italic_e ( 2 ) and contour enhancement via invertible orientation scores part ii: Nonlinear left-invariant diffusions on invertible orientation scores. Quarterly of applied mathematics, 68(2):293–331, 2010b.
  20. On the universality of rotation equivariant point cloud networks. In International Conference on Learning Representations, 2020.
  21. Fast graph representation learning with PyTorch Geometric. In ICLR Workshop on Representation Learning on Graphs and Manifolds, 2019.
  22. Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. In International Conference on Machine Learning, pp.  3165–3176. PMLR, 2020.
  23. Se (3)-transformers: 3d roto-translation equivariant attention networks. Advances in neural information processing systems, 33:1970–1981, 2020.
  24. E (n) equivariant normalizing flows. Advances in Neural Information Processing Systems, 34:4181–4192, 2021.
  25. Directional message passing for molecular graphs. In International Conference on Learning Representations, 2019.
  26. Gemnet: Universal directional graph neural networks for molecules. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=HS_sOaxS9K-.
  27. Symmetry-adapted generation of 3d point sets for the targeted discovery of molecules. Advances in neural information processing systems, 32, 2019.
  28. e3nn: Euclidean neural networks, 2022. URL https://arxiv.org/abs/2207.09453.
  29. Neural message passing for quantum chemistry. In International conference on machine learning, pp.  1263–1272. PMLR, 2017a.
  30. Neural message passing for quantum chemistry. In Doina Precup and Yee Whye Teh (eds.), Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pp.  1263–1272. PMLR, 06–11 Aug 2017b. URL https://proceedings.mlr.press/v70/gilmer17a.html.
  31. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840–6851, 2020.
  32. Equivariant diffusion for molecule generation in 3d. In International conference on machine learning, pp.  8867–8887. PMLR, 2022.
  33. Design and processing of invertible orientation scores of 3d images. Journal of mathematical imaging and vision, 60:1427–1458, 2018.
  34. Variational diffusion models. Advances in neural information processing systems, 34:21696–21707, 2021.
  35. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  36. Neural relational inference for interacting systems. In International conference on machine learning, pp.  2688–2697. PMLR, 2018.
  37. Exploiting redundancy: Separable group convolutional networks on lie groups. In International Conference on Machine Learning, pp.  11359–11386. PMLR, 2022.
  38. Equivariant flows: exact likelihood generative learning for symmetric densities. In International conference on machine learning, pp.  5361–5370. PMLR, 2020.
  39. An exploration of conditioning methods in graph neural networks. In ICLR 2023-Machine Learning for Drug Discovery workshop, 2023.
  40. On the generalization of equivariance and convolution in neural networks to the action of compact groups. In International Conference on Machine Learning, pp.  2747–2755. PMLR, 2018.
  41. Clebsch–gordan nets: a fully fourier space spherical convolutional neural network. Advances in Neural Information Processing Systems, 31, 2018.
  42. Regular se (3) group convolutions for volumetric medical image analysis. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2023. arXiv preprint arXiv:2306.13960, 2023.
  43. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  44. Spherical message passing for 3d molecular graphs. In International Conference on Learning Representations, 2022a. URL https://openreview.net/forum?id=givsRXsOt9r.
  45. A convnet for the 2020s. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  11976–11986, 2022b.
  46. Learning local equivariant representations for large-scale atomistic dynamics. Nature Communications, 14(1):579, 2023.
  47. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
  48. Hyena hierarchy: Towards larger convolutional language models. arXiv preprint arXiv:2302.10866, 2023.
  49. New approximation of a scale space kernel on se (3) and applications in neuroimaging. In International Conference on Scale Space and Variational Methods in Computer Vision, pp.  40–52. Springer, 2015.
  50. Informing geometric deep learning with electronic interactions to accelerate quantum chemistry. Proceedings of the National Academy of Sciences, 119(31):e2205221119, 2022.
  51. Rupp M.and Von Lilienfeld O.A. Ramakrishnan R., Dral P.O. Quantum chemistry structures and properties of 134 kilo molecules, 2014.
  52. Ckconv: Continuous kernel convolution for sequential data. In International Conference on Learning Representations, 2021.
  53. Clifford group equivariant neural networks. arXiv preprint arXiv:2305.11141, 2023.
  54. E (n) equivariant graph neural networks. In International conference on machine learning, pp.  9323–9332. PMLR, 2021.
  55. Schnetpack 2.0: A neural network toolbox for atomistic machine learning. The Journal of Chemical Physics, 158(14), 2023.
  56. Symmetry-aware actor-critic for 3d molecular design. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=jEYKjPE1xYN.
  57. Graphvae: Towards generation of small graphs using variational autoencoders. In Artificial Neural Networks and Machine Learning–ICANN 2018: 27th International Conference on Artificial Neural Networks, Rhodes, Greece, October 4-7, 2018, Proceedings, Part I 27, pp.  412–422. Springer, 2018.
  58. Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp.  2256–2265. PMLR, 2015.
  59. Maximum likelihood training of score-based diffusion models. Advances in Neural Information Processing Systems, 34:1415–1428, 2021.
  60. Scale-equivariant steerable networks. In International Conference on Learning Representations, 2019.
  61. Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, 33:7537–7547, 2020.
  62. Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds. arXiv preprint arXiv:1802.08219, 2018.
  63. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  64. Midi: Mixed graph and 3d denoising diffusion for molecule generation. arXiv preprint arXiv:2302.09048, 2023.
  65. Scalars are universal: Equivariant machine learning, structured like classical physics. Advances in Neural Information Processing Systems, 34:28848–28863, 2021.
  66. Visnet: a scalable and accurate geometric deep learning potential for molecular dynamics simulation. arXiv preprint arXiv:2210.16518, 2022.
  67. General e (2)-equivariant steerable cnns. Advances in neural information processing systems, 32, 2019.
  68. Learning steerable filters for rotation equivariant cnns. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  849–858, 2018.
  69. Coordinate independent convolutional networks–isometry and gauge equivariant convolutions on riemannian manifolds. arXiv preprint arXiv:2106.06020, 2021.
  70. Equivariant and Coordinate Independent Convolutional Networks. 2023. URL https://maurice-weiler.gitlab.io/cnn_book/EquivariantAndCoordinateIndependentCNNs.pdf.
  71. Cubenet: Equivariance to 3d rotation and translation. In Proceedings of the European Conference on Computer Vision (ECCV), pp.  567–584, 2018.
  72. Geodiff: A geometric diffusion model for molecular conformation generation. arXiv preprint arXiv:2203.02923, 2022.
Citations (14)

Summary

  • The paper introduces a novel weight-sharing approach using invariant attributes to design efficient SE(n) equivariant neural networks.
  • It formalizes group convolutions in position-orientation space to reduce computational complexity while capturing isotropic and directional features.
  • Empirical validations on benchmarks such as molecular generation and trajectory forecasting underscore its state-of-the-art performance.

An Analysis of SE(n) Equivariant Networks via Weight-Sharing in Position-Orientation Space

The paper "Fast, Expressive SE(n)(n) Equivariant Networks through Weight-Sharing in Position-Orientation Space" introduces a fresh approach to designing SE(n) equivariant neural networks, focusing on 3D point cloud processing. It employs the concept of weight-sharing grounded in geometric attributes, optimized for group convolutions within the position-orientation space. By formalizing weight sharing as a function of message-passing operations conditioned on invariant attributes, the paper provides a structured methodology to construct efficient and expressive group convolutional networks.

Invariant Attributes and Weight-Sharing

The authors establish an innovative framework to handle geometric data by identifying equivalence classes of point-pairs invariant to transformations in the SE(n) group. This is achieved by employing geometrically optimal edge attributes derived through the theory of homogeneous spaces. The paper thoroughly analyzes the SE(n) group action on equivalence classes and induces a parameterization involving invariant attributes. These attributes act as a means to uniquely identify equivalence classes and facilitate weight-sharing within the neural architecture.

The research articulates specific mappings for various homogeneous spaces associated with SE(n), such as position space Rn\mathbb{R}^n, the hybrid position-orientation space R3×S2\mathbb{R}^3 \times S^2, and the full group space SE(3)SE(3). These mappings ensure a bijective correspondence with the space of all equivalence classes, allowing a robust foundation for weight-sharing implementations across different space representations. The theoretical grounding of these mappings enhances the understanding of how invariant attributes contribute to the full expressivity and efficiency of equivariant layers.

Efficient Architecture Design

The practical application of their theoretical insights manifests in the development of a separable group convolutional framework, tailored for R3×S2\mathbb{R}^3 \times S^2. This approach promises computational efficiency by partitioning group convolution into spatial, spherical, and channel mixing operations. By leveraging invariant attributes, the architecture reduces computational complexity without compromising on capturing intricate isotropic and directional features within point clouds.

The architecture constructs equivalence classes by implementing an effector mechanism through invariant attributes in message functions within the deep learning framework. This design choice facilitates processing equivalently under graph permutations and transformations, aligning with equivariant inductive biases intrinsic to the SE(n) group.

Empirical Validation

The paper demonstrates state-of-the-art performance across several benchmarks, notably in interatomic potential energy prediction, trajectory forecasting in N-body systems, and molecular generation tasks. The performance efficiency and accuracy underscore the potential of refining deep learning models with a strategic focus on group equivariance through geometrically aligned invariant mappings.

Implications and Future Directions

This research illuminates the prospect of enhancing neural architectures by embedding geometric and algebraic properties through invariant mappings. While its immediate application in 3D data processing is clear, the structured view of weight-sharing and convolutions furnished by invariant attributes paves the way for extending this approach to broader machine learning challenges, particularly those requiring efficiency and expressivity in high-dimensional and group-transformed data spaces.

Investigating the potential of expanding to dynamic scenarios where the group action may involve more complex transformations, or exploring other groups and homogeneous spaces, embodies plausible future research trajectories. Moreover, scaling the theoretical framework to accommodate multi-modal data and integrating with other neural network paradigms such as attention mechanisms could form innovative future explorations.

In conclusion, the paper's methodology in deriving weight-sharing schemes via invariant attributes conspicuously integrates theoretical and practical facets of deep learning, presenting a comprehensive toolkit for next-generation SE(n) equivariant networks.