Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Clifford-Steerable Convolutional Neural Networks (2402.14730v3)

Published 22 Feb 2024 in cs.LG and cs.AI

Abstract: We present Clifford-Steerable Convolutional Neural Networks (CS-CNNs), a novel class of $\mathrm{E}(p, q)$-equivariant CNNs. CS-CNNs process multivector fields on pseudo-Euclidean spaces $\mathbb{R}{p,q}$. They cover, for instance, $\mathrm{E}(3)$-equivariance on $\mathbb{R}3$ and Poincar\'e-equivariance on Minkowski spacetime $\mathbb{R}{1,3}$. Our approach is based on an implicit parametrization of $\mathrm{O}(p,q)$-steerable kernels via Clifford group equivariant neural networks. We significantly and consistently outperform baseline methods on fluid dynamics as well as relativistic electrodynamics forecasting tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Mace: Higher order equivariant message passing neural networks for fast and accurate force fields. Conference on Neural Information Processing Systems (NeurIPS), 2022.
  2. Bekkers, E. B-spline CNNs on Lie groups. International Conference on Learning Representations (ICLR), 2020.
  3. Roto-translation covariant convolutional networks for medical image analysis. International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2018.
  4. Clifford neural layers for PDE modeling. International Conference on Learning Representations (ICLR), 2023.
  5. Geometric algebra transformers. Conference on Neural Information Processing Systems (NeurIPS), 2023.
  6. A program to build E(N)-equivariant steerable CNNs. International Conference on Learning Representations (ICLR), 2022.
  7. Group equivariant convolutional networks. International Conference on Machine Learning (ICML), 2016.
  8. Steerable CNNs. International Conference on Learning Representations (ICLR), 2017.
  9. A general theory of equivariant CNNs on homogeneous spaces. Conference on Neural Information Processing Systems (NeurIPS), 2019a.
  10. Gauge Equivariant Convolutional Networks and the Icosahedral CNN. International Conference on Machine Learning (ICML), 2019b.
  11. Gauge Equivariant Mesh CNNs: Anisotropic convolutions on geometric graphs. International Conference on Learning Representations (ICLR), 2021.
  12. Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. International Conference on Machine Learning (ICML), 2020.
  13. A practical method for constructing equivariant multilayer perceptrons for arbitrary matrix groups. In International conference on machine learning, pp.  3318–3328. PMLR, 2021.
  14. Euclidean neural networks: e3nn. Zenodo. https://doi. org/10.5281/zenodo, 2020.
  15. Scale steerable filters for locally scale-invariant convolutional neural networks. arXiv preprint arXiv:1906.03861, 2019.
  16. Towards multi-spatiotemporal-scale generalized pde modeling. arXiv preprint arXiv:2209.15616, 2022.
  17. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415, 2016.
  18. Hestenes, D. Multivector calculus. J. Math. Anal. Appl, 24(2):313–325, 1968.
  19. Hestenes, D. Space-time algebra. Springer, 2015.
  20. Hitzer, E. M. Multivector differential calculus. Advances in Applied Clifford Algebras, 12:135–182, 2002.
  21. Learning to control pdes with differentiable physics. International Conference on Learning Representations (ICLR), 2020.
  22. Steerable Partial Differential Operators for Equivariant Neural Networks. International Conference on Learning Representations (ICLR), 2022.
  23. Adam: A method for stochastic optimization. International Conference on Learning Representations (ICLR), 2014.
  24. A Wigner-Eckart Theorem for Group Equivariant Convolution Kernels. International Conference on Learning Representations (ICLR), 2020.
  25. A multivector derivative approach to lagrangian field theory. Foundations of Physics, 23(10):1295–1327, 1993.
  26. Lindeberg, T. Scale-space. 2009.
  27. Clifford group equivariant simplicial message passing networks. arXiv preprint arXiv:2402.10011, 2024.
  28. SGDR: Stochastic gradient descent with warm restarts. International Conference on Learning Representations (ICLR), 2017.
  29. Scale equivariance in CNNs with vector fields. arXiv preprint arXiv:1807.11783, 2018.
  30. Dimensional scaffolding of electromagnetism using geometric algebra. European Journal of Physics, 42, 2020. URL https://api.semanticscholar.org/CorpusID:215754421.
  31. Wavelet networks: Scale equivariant learning from raw waveforms. arXiv preprint arXiv:2006.05259, 2020.
  32. Clifford Group Equivariant Neural Networks. Conference on Neural Information Processing Systems (NeurIPS), 2023a.
  33. Geometric clifford algebra networks. International Conference on Machine Learning (ICML), 2023b.
  34. Computing representations for lie algebraic networks. NeurIPS 2022 Workshop on Symmetry and Geometry in Neural Representations, 2022.
  35. Scale-equivariant steerable networks. International Conference on Learning Representations (ICLR), 2020.
  36. Computational electromagnetics: the finite-difference time-domain method. The Electrical Engineering Handbook, 3(629-670):15, 2005.
  37. Incorporating symmetry into deep dynamics models for improved generalization. International Conference on Learning Representations (ICLR), 2020.
  38. General E(2)-Equivariant Steerable CNNs. Conference on Neural Information Processing Systems (NeurIPS), 2019.
  39. 3d steerable cnns: Learning rotationally equivariant features in volumetric data. Advances in Neural Information Processing Systems (NeurIPS), 2018a.
  40. Learning Steerable Filters for Rotation Equivariant CNNs. Conference on Computer Vision and Pattern Recognition (CVPR), 2018b.
  41. Coordinate Independent Convolutional Networks – Isometry and Gauge Equivariant Convolutions on Riemannian Manifolds. arXiv preprint arXiv:2106.06020, 2021.
  42. Equivariant and Coordinate Independent Convolutional Networks. 2023. URL https://maurice-weiler.gitlab.io/cnn_book/EquivariantAndCoordinateIndependentCNNs.pdf.
  43. Deep scale-spaces: Equivariance over scale. Conference on Neural Information Processing Systems (NeurIPS), 2019.
  44. Group normalization. Proceedings of the European conference on computer vision (ECCV), 2018.
  45. Similarity equivariant linear transformation of joint orientation-scale space representations. arXiv preprint arXiv:2203.06786, 2022.
  46. Implicit neural convolutional kernels for steerable cnns. Conference on Neural Information Processing Systems (NeurIPS), 2022.
  47. Scale-equivariant neural networks with decomposed convolutional filters. arXiv preprint arXiv:1909.11193, 2019.
Citations (9)

Summary

  • The paper introduces CS-CNNs that extend steerable CNN frameworks to pseudo-Euclidean spaces, ensuring equivariance to complex symmetry groups.
  • The paper utilizes Clifford algebra and implicit parameterization via equivariant MLPs to overcome the limitations of conventional analytical methods.
  • The paper demonstrates robust empirical gains over baselines in tasks such as fluid dynamics and electrodynamics, highlighting significant sample efficiency.

Clifford-Steerable Convolutional Neural Networks

The paper presents a novel class of Convolutional Neural Networks (CNNs) termed Clifford-Steerable Convolutional Neural Networks (CS-CNNs). These networks are designed to process multivector fields on pseudo-Euclidean spaces denoted by Rp,qR^{p,q}. They offer a notable advancement by being equivariant to the symmetries in these spaces, such as (n)(n)-equivariance on RnR^n and Poincaré-equivariance on Minkowski spacetime R1,3R^{1,3}—addressing the long-standing challenge of designing CNNs that respect these symmetries.

Equivariance and Symmetry

The central premise of the paper hinges on the concept of equivariance, which is critical for ensuring that a neural network's predictions respect the inherent symmetries present in the data. In physical systems, this often translates to the requirement that the underlying differential equations or the predictor respect transformation properties like rotations (isometries) in Euclidean space R3R^3 or Lorentz transformations in Minkowski spacetime R1,3R^{1,3}.

CS-CNNs extend the framework of steerable CNNs, previously limited to Euclidean spaces and compact groups, to pseudo-Euclidean spaces with mixed signatures. This generalization is non-trivial due to the non-compact nature of groups like (p,q)(p,q) which necessitate novel approaches both in theory and implementation.

Clifford Algebra and Kernel Parametrization

The core of these networks is their reliance on Clifford algebra to achieve equivariance. The Clifford algebra offers a rich mathematical structure that can naturally incorporate the geometric properties of the base space. For an input vector space V=Rp,qV = R^{p,q}, the Clifford algebra (V,η)(V, \eta) encompasses vectors, scalars, and higher-grade multivectors, providing a versatile feature space for deep learning models.

The steerable kernels in CS-CNNs are implicitly parameterized via neural networks that are equivariant to the Clifford group. This implicit parameterization approach overcomes the limitations of analytical solutions which would be cumbersome and highly specialized for each group structure. Instead, the implicit method leverages equivariant multi-layer perceptrons (MLPs) to satisfy the equivariance constraints inherently.

Achievements and Results

CS-CNNs are evaluated on tasks such as fluid dynamics forecasting on R2R^2, and electrodynamics simulations on R3R^3 and R1,2R^{1,2}. They consistently outperform baseline methods, including conventional steerable CNNs and non-equivariant Clifford CNNs. This performance is robust across various dataset sizes, demonstrating significant sample efficiency. In particular, CS-CNNs retain full symmetry properties while providing substantial empirical gains.

Practical and Theoretical Implications

The implications of CS-CNNs extend both practically and theoretically. Practically, they offer improved prediction capabilities and sample efficiency for complex physical systems like fluid dynamics and electrodynamics, which are described by partial differential equations invariant under certain symmetries. Theoretically, CS-CNNs broaden the scope of equivariant deep learning models to incorporate pseudo-Euclidean symmetries, a leap from the previously explored Euclidean and compact Lie group settings.

Future Directions

Future developments could focus on the generalizations to other types of geometric features and more intricate symmetries, potentially even beyond pseudo-Euclidean spaces. Additionally, the models could be further optimized for specific applications in physics and engineering, such as better integration with numerical solvers for PDEs. Another promising direction is the exploration of CS-CNNs in more generalized settings, such as Riemannian and pseudo-Riemannian manifolds, enhancing their capability to handle a wider range of physical systems.

Conclusion

CS-CNNs represent a significant advancement in the domain of geometric deep learning by integrating Clifford algebra with the framework of steerable CNNs. Their ability to handle the complex symmetries of pseudo-Euclidean spaces while ensuring computational efficiency and robust performance underlines their potential both as a research tool and in practical applications. This work opens new avenues for the exploration and application of equivariant neural networks to model intricate physical systems accurately and efficiently.

HackerNews