Papers
Topics
Authors
Recent
2000 character limit reached

E(n) Equivariant Topological Neural Networks

Published 24 May 2024 in cs.LG and cs.NE | (2405.15429v5)

Abstract: Graph neural networks excel at modeling pairwise interactions, but they cannot flexibly accommodate higher-order interactions and features. Topological deep learning (TDL) has emerged recently as a promising tool for addressing this issue. TDL enables the principled modeling of arbitrary multi-way, hierarchical higher-order interactions by operating on combinatorial topological spaces, such as simplicial or cell complexes, instead of graphs. However, little is known about how to leverage geometric features such as positions and velocities for TDL. This paper introduces E(n)-Equivariant Topological Neural Networks (ETNNs), which are E(n)-equivariant message-passing networks operating on combinatorial complexes, formal objects unifying graphs, hypergraphs, simplicial, path, and cell complexes. ETNNs incorporate geometric node features while respecting rotation, reflection, and translation equivariance. Moreover, being TDL models, ETNNs are natively ready for settings with heterogeneous interactions. We provide a theoretical analysis to show the improved expressiveness of ETNNs over architectures for geometric graphs. We also show how E(n)-equivariant variants of TDL models can be directly derived from our framework. The broad applicability of ETNNs is demonstrated through two tasks of vastly different scales: i) molecular property prediction on the QM9 benchmark and ii) land-use regression for hyper-local estimation of air pollution with multi-resolution irregular geospatial data. The results indicate that ETNNs are an effective tool for learning from diverse types of richly structured data, as they match or surpass SotA equivariant TDL models with a significantly smaller computational burden, thus highlighting the benefits of a principled geometric inductive bias. Our implementation of ETNNs can be found at https://github.com/NSAPH-Projects/topological-equivariant-networks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (107)
  1. “Neural message passing for quantum chemistry,” in International Conference on Machine learning, 2017.
  2. “Graph neural networks in particle physics,” Machine Learning: Science and Technology, vol. 2, no. 2, pp. 021001, dec 2020.
  3. “Deepis: Susceptibility estimation on social networks,” in Proceedings of the 14th ACM International Conference on Web Search and Data Mining, New York, NY, USA, 2021, WSDM ’21, p. 761–769, Association for Computing Machinery.
  4. “Spectral networks and locally connected networks on graphs,” in International Conference on Learning Representations, Banff, Canada, 2014.
  5. “A new model for learning in graph domains,” in IEEE International Joint Conference on Neural Networks, 2005.
  6. “Neural message passing for quantum chemistry,” in Proceedings of the 34th International Conference on Machine Learning - Volume 70. 2017, ICML’17, p. 1263–1272, JMLR.org.
  7. T. N. Kipf and M. Welling, “Semi-Supervised Classification with Graph Convolutional Networks,” in International Conference on Learning Representations, 2017.
  8. M. Zhang and Y. Chen, “Link prediction based on graph neural networks,” Advances in Neural Information Processing Systems, 2018.
  9. “Highly accurate protein structure prediction with alphafold,” Nature, 2021.
  10. S. Barbarossa and S. Sardellitti, “Topological signal processing over simplicial complexes,” IEEE Transactions on Signal Processing, 2020.
  11. “Topological deep learning: Going beyond graph data,” 2023.
  12. “Position paper: Challenges and opportunities in topological deep learning,” arXiv preprint arXiv:2402.08871, 2024.
  13. Discrete calculus: Applied analysis on graphs for computational science, Springer, 2010.
  14. Cristian Bodnar, Topological Deep Learning: Graphs, Complexes, Sheaves, Ph.D. thesis, Cambridge University, 2023.
  15. Claudio Battiloro, Signal Processing and Learning over Topological Spaces, Ph.D. thesis, Sapienza Unviersity of Rome, 2024.
  16. “Weisfeiler and lehman go topological: Message passing simplicial networks,” in International Conference on Machine Learning, 2021.
  17. “Generalized simplicial attention neural networks,” arXiv preprint arXiv:2309.02138, 2023.
  18. “Weisfeiler and lehman go cellular: Cw networks,” in Advances in Neural Information Processing Systems, 2021.
  19. “Weisfeiler and lehman go paths: Learning topological features via path complexes,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2024, vol. 38, pp. 15382–15391.
  20. “Expressivity-preserving GNN simulation,” in Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  21. “How powerful are graph neural networks?,” in International Conference on Learning Representations, 2019.
  22. “Cin++: Enhancing topological message passing,” arXiv preprint arXiv:2306.03561, 2023.
  23. “Cell attention networks,” in 2023 International Joint Conference on Neural Networks (IJCNN). IEEE, 2023, pp. 1–8.
  24. “Neural sheaf diffusion: A topological perspective on heterophily and oversmoothing in GNNs,” in Advances in Neural Information Processing Systems, 2022.
  25. “Sheaf neural networks,” 2020.
  26. “E (n) equivariant graph neural networks,” in International conference on machine learning. PMLR, 2021, pp. 9323–9332.
  27. “E (n)𝑛(n)( italic_n ) equivariant message passing simplicial networks,” in International Conference on Machine Learning. PMLR, 2023, pp. 9071–9081.
  28. “Geometrically equivariant graph neural networks: A survey,” arXiv preprint arXiv:2202.07230, 2022.
  29. “Topological signal processing over cell complexes,” in Asilomar Conference on Signals, Systems, and Computers, 2021.
  30. “Geometric deep learning: Grids, groups, graphs, geodesics, and gauges,” arXiv preprint arXiv:2104.13478, 2021.
  31. “On the usage of average hausdorff distance for segmentation performance assessment: hidden error when used for ranking,” European Radiology Experimental, vol. 5, no. 1, Jan. 2021.
  32. “Between shapes, using the hausdorff distance,” Computational Geometry, vol. 100, pp. 101817, 2022.
  33. “On the expressive power of geometric graph neural networks,” in International Conference on Machine Learning. PMLR, 2023, pp. 15330–15355.
  34. “Multiscale statistical models for hierarchical spatial aggregation,” Geographical Analysis, vol. 33, no. 2, pp. 95–118, 2001.
  35. “African soil properties and nutrients mapped at 30 m spatial resolution using two-scale ensemble machine learning,” Scientific Reports, vol. 11, no. 1, Mar. 2021.
  36. “Quantum chemistry structures and properties of 134 kilo molecules,” Scientific Data, vol. 1, no. 1, Aug. 2014.
  37. “Vn-egnn: E (3)-equivariant graph neural networks with virtual nodes enhance protein binding site identification,” arXiv preprint arXiv:2404.07194, 2024.
  38. “Hyperlocal environmental data with a mobile platform in urban environments,” Scientific data, vol. 10, no. 1, pp. 524, 2023.
  39. “Clifford group equivariant simplicial message passing networks,” in The Twelfth International Conference on Learning Representations, 2024.
  40. “Heterogeneous graph neural network,” in Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, 2019, pp. 793–803.
  41. “Simplicial representation learning with neural $k$-forms,” in The Twelfth International Conference on Learning Representations, 2024.
  42. “Signal processing on higher-order networks: Livin’on the edge… and beyond,” Signal Processing, 2021.
  43. “Signal processing on cell complexes,” in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 8852–8856.
  44. “Simplicial neural networks,” in Advances in Neural Information Processing Systems Workshop on Topological Data Analysis and Beyond, 2020.
  45. “Efficient representation learning for higher-order data with simplicial complexes,” in Proceedings of the First Learning on Graphs Conference, Bastian Rieck and Razvan Pascanu, Eds. 09–12 Dec 2022, vol. 198 of Proceedings of Machine Learning Research, pp. 13:1–13:21, PMLR.
  46. “Cell complex neural networks,” in Advances in Neural Information Processing Systems Workshop on TDA & Beyond, 2020.
  47. “Convolutional learning on simplicial complexes,” 2023.
  48. “Principled simplicial neural networks for trajectory prediction,” in International Conference on Machine Learning, 2021.
  49. “Higher-order attention networks,” 2022.
  50. “Simplicial attention neural networks,” 2022.
  51. “Simplicial attention networks,” in International Conference on Learning Representations Workshop on Geometrical and Topological Representation Learning, 2022.
  52. J. Hansen and R. Ghrist, “Toward a spectral theory of cellular sheaves,” Journal of Applied and Computational Topology, 2019.
  53. “Tangent bundle filters and neural networks: From manifolds to cellular sheaves and back,” in ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023, pp. 1–5.
  54. “Tangent bundle convolutional learning: from manifolds to cellular sheaves and back,” IEEE Transactions on Signal Processing, 2024.
  55. “Sheaf neural networks with connection laplacians,” 2022.
  56. “SaNN: Simple yet powerful simplicial-aware neural networks,” in The Twelfth International Conference on Learning Representations, 2024.
  57. “Hypergraph-mlp: learning on hypergraphs without message passing,” in ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2024, pp. 13476–13480.
  58. “From latent graph to latent topology inference: Differentiable cell complex module,” in The Twelfth International Conference on Learning Representations, 2024.
  59. “Architectures of topological deep learning: A survey on topological neural networks,” 2023.
  60. “Deep residual learning for image recognition,” in IEEE conference on computer vision and pattern recognition, 2016.
  61. “Group equivariant convolutional networks,” in International conference on machine learning. PMLR, 2016, pp. 2990–2999.
  62. Equivariant and Coordinate Independent Convolutional Networks, 2023.
  63. “Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds,” arXiv preprint arXiv:1802.08219, 2018.
  64. “Se (3)-transformers: 3d roto-translation equivariant attention networks,” Advances in neural information processing systems, vol. 33, pp. 1970–1981, 2020.
  65. “A general theory of equivariant cnns on homogeneous spaces,” Advances in neural information processing systems, vol. 32, 2019.
  66. “Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data,” in International Conference on Machine Learning. PMLR, 2020, pp. 3165–3176.
  67. “Lietransformer: Equivariant self-attention for lie groups,” in International Conference on Machine Learning. PMLR, 2021, pp. 4533–4543.
  68. “Equivariant flows: sampling configurations for multi-body systems with symmetric energies,” arXiv preprint arXiv:1910.00753, 2019.
  69. “Schnet: A continuous-filter convolutional neural network for modeling quantum interactions,” Advances in neural information processing systems, vol. 30, 2017.
  70. “Geometric and physical quantities improve e (3) equivariant message passing,” arXiv preprint arXiv:2110.02905, 2021.
  71. “E (3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials,” Nature communications, vol. 13, no. 1, pp. 2453, 2022.
  72. “A characterization theorem for equivariant networks with point-wise activations,” in The Twelfth International Conference on Learning Representations, 2024.
  73. “Clifford group equivariant neural networks,” Advances in Neural Information Processing Systems, vol. 36, 2024.
  74. “Geometric algebra transformer,” Advances in Neural Information Processing Systems, vol. 36, 2024.
  75. “Two’s company, three (or more) is a simplex,” Journal of computational neuroscience, 2016.
  76. “A topological representation of branching neuronal morphologies,” Neuroinfor., vol. 16, no. 1, pp. 3–13, 2018.
  77. “The shape of collaborations,” EPJ Data Sci., vol. 6, no. 1, pp. 18, 2017.
  78. L. Lim, “Hodge Laplacians on graphs,” Siam Review, 2020.
  79. D. Mulder and G. Bianconi, “Network geometry and complexity,” Journal of Statistical Physics, vol. 173, no. 3, pp. 783–805, 2018.
  80. “Least squares ranking on graphs, Hodge Laplacians, time optimality, and iterative methods,” arXiv preprint arXiv:1011.1716, 2010.
  81. “Recent developments in hodge theory: a discussion of techniques and results,” 1975.
  82. “Topological signal processing over weighted simplicial complexes,” in ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023, pp. 1–5.
  83. “Graph and hodge laplacians: Similarity and difference,” arXiv preprint arXiv:2204.12218, 2022.
  84. “Hypergraph neural networks,” in Proceedings of the AAAI conference on artificial intelligence, 2019, vol. 33, pp. 3558–3565.
  85. “Message passing neural networks for hypergraphs,” in International Conference on Artificial Neural Networks. Springer, 2022, pp. 583–592.
  86. B. Weisfeiler and A. Leman, “The reduction of a graph to canonical form and the algebra which appears therein,” NTI, Series, 1968.
  87. “The relationship between aristotelian and hasse diagrams,” in International Conference on Theory and Application of Diagrams. Springer, 2014, pp. 213–227.
  88. “On over-squashing in message passing neural networks: The impact of width, depth, and topology,” in International Conference on Machine Learning. PMLR, 2023, pp. 7865–7885.
  89. “Directional message passing for molecular graphs,” in International Conference on Learning Representations, 2020.
  90. “Could graph neural networks learn better molecular representation for drug discovery? a comparison study of descriptor-based and graph-based models,” Journal of Cheminformatics, vol. 13, no. 1, Feb. 2021.
  91. “Multi-view graph neural networks for molecular property prediction,” 2020.
  92. “Protein representation learning by geometric structure pretraining,” in The Eleventh International Conference on Learning Representations, 2023.
  93. “Neural atoms: Propagating long-range interaction in molecular graphs through efficient communication channel,” in The Twelfth International Conference on Learning Representations, 2024.
  94. “Molecular hypergraph neural networks,” arXiv preprint arXiv:2312.13136, 2023.
  95. “A molecular hypergraph convolutional network with functional group information,” 2021.
  96. “A spatial–temporal hypergraph based method for service recommendation in the mobile internet of things-enabled service platform,” Advanced Engineering Informatics, vol. 57, pp. 102038, 2023.
  97. “Unveiling patterns of international communities in a global city using mobile phone data,” EPJ Data Science, vol. 4, no. 1, Apr. 2015.
  98. “Modeling the spread of the zika virus using topological data analysis,” PLOS ONE, vol. 13, no. 2, pp. e0192120, Feb. 2018.
  99. “The topological 'shape' of brexit,” SSRN Electronic Journal, 2016.
  100. “Persistent homology of geospatial data: A case study with voting,” SIAM Review, vol. 63, no. 1, pp. 67–99, 2021.
  101. “Analysis of spatial and spatiotemporal anomalies using persistent homology: Case studies with covid-19 data,” SIAM Journal on Mathematics of Data Science, vol. 4, no. 3, pp. 1116–1144, 2022.
  102. Cormorant: covariant molecular neural networks, Curran Associates Inc., Red Hook, NY, USA, 2019.
  103. “Long-term mortality burden trends attributed to black carbon and pm2· 5 from wildfire emissions across the continental usa from 2000 to 2020: a deep learning modelling study,” The Lancet Planetary Health, vol. 7, no. 12, pp. e963–e975, 2023.
  104. New York City Department of City Planning, “Primary land use tax lot output (pluto) data,” https://www1.nyc.gov/site/planning/data-maps/open-data/dwn-pluto-mappluto.page, 2024, Accessed: 2024-05-22.
  105. New York City Department of Transportation, “Average daily traffic (aadt) data,” https://data.cityofnewyork.us/Transportation/Average-Daily-Traffic-AADT-/fp86-8v8r, 2024, Accessed: 2024-05-22.
  106. Geoff Boeing, “Osmnx: A python package to work with openstreetmap data,” https://github.com/gboeing/osmnx, 2017, Version 1.1.1, Accessed: 2024-05-22.
  107. Peter J. Huber, “Robust estimation of a location parameter,” Annals of Mathematical Statistics, vol. 35, no. 1, pp. 73–101, 1964.
Citations (3)

Summary

  • The paper proposes a novel ETNN architecture that integrates geometric features with E(n) equivariance to model higher-order interactions in combinatorial complexes.
  • It demonstrates broad applicability by excelling in tasks like molecular property prediction on QM9 and fine-grained air pollution downscaling using heterogeneous geospatial data.
  • ETNNs efficiently compute geometric invariants such as pairwise and Hausdorff distances, setting a new benchmark in expressive, symmetry-preserving deep learning.

E(n) Equivariant Topological Neural Networks

Graph neural networks (GNNs) have gained prominence for their ability to address various applications involving graph-structured data, including computational chemistry, physics simulations, and social networks. However, the inherent pairwise nature of graphs restricts GNNs from effectively modeling higher-order interactions and features. Topological deep learning (TDL) endeavors to overcome these limitations by introducing more sophisticated structures such as simplicial and cell complexes. Despite the progress, the integration and leveraging of geometric features in TDL models have remained largely unexplored. The paper "E(n) Equivariant Topological Neural Networks (ETNNs)" proposes a novel architecture designed to address this gap.

Overview of ETNNs

ETNNs are message-passing networks operating over combinatorial complexes. These complexes generalize various combinatorial topological spaces, including graphs, hypergraphs, and higher structures such as simplicial and cell complexes. ETNNs incorporate geometric node features while ensuring E(n) group equivariance, which encompasses rotations and translations. By doing so, ETNNs cater to scenarios requiring heterogeneous interaction modeling, providing a more expressive framework compared to traditional GNNs.

Key Contributions

  1. Expressive Power: The paper demonstrates that ETNNs surpass the expressiveness of existing architectures for geometric graphs, such as EGNNs, by showing how they can model complex hierarchical interactions in a principled manner. This is achieved through the use of geometric invariants and scalarization techniques, which maintain E(n) equivariance.
  2. Broad Applicability: The authors extend ETNNs to a variety of practical tasks. The experiments highlight two distinct applications:
    • Molecular property prediction using the QM9 benchmark, which underscores the capability of ETNNs to handle richly structured molecular data.
    • Land-use regression for fine-grained air pollution estimation using heterogeneous geospatial data, demonstrating ETNNs' utility in integrating multi-resolution datasets.
  3. Geometric Feature Integration: ETNNs exploit various geometric invariants, including pairwise distances, Hausdorff distances, and volumes of convex hulls. These invariants are computed efficiently and used to update non-geometric and geometric features of nodes within the complexes.

Experimental Results

  1. Molecular Property Prediction: The evaluation on the QM9 dataset reveals that ETNNs outperform traditional methods like EGNN on multiple molecular properties. Notably, ETNNs achieve a new state-of-the-art performance for the dipole moment (μ\mu), with a reduction in the mean absolute error over prior models.
  2. Air Pollution Downscaling: The paper introduces a new benchmark for hyper-local air pollution prediction using multi-resolution irregular geospatial data. ETNNs surpass baselines like MLP, GNN, and EGNN, demonstrating their effectiveness in handling complex heterogeneous data structures. The ablation studies further underscore the contributions of various components, such as the inclusion of geometric features and the presence of virtual cells.

Implications and Future Directions

The theoretical and practical advancements presented by ETNNs open several avenues for future research:

  1. Extensions to Other Equivariances: Beyond E(n) equivariance, exploring symmetries for other groups or product groups could extend the applicability of ETNNs to more complex domains.
  2. Dynamic Scenarios: ETNNs are currently designed for static settings. Future work could extend them to dynamic settings, enabling modeling of time-varying data and interactions.
  3. Scalability Improvements: While ETNNs show promise for moderately sized datasets, scalability remains a challenge for very large datasets. Techniques like neighbor sampling and message-passing-free architectures could be explored for efficiency.
  4. Unsupervised Learning: Developing unsupervised or self-supervised versions of ETNNs could enhance their ability to learn meaningful representations from large, unlabeled datasets.

Conclusion

E(n) Equivariant Topological Neural Networks represent a significant advancement in topological deep learning by offering a more expressive and flexible way to model higher-order, hierarchical interactions while respecting geometric symmetries. Their application to both molecular data and complex geospatial data sets showcases their versatility and capability. The paper provides a solid theoretical foundation and demonstrates practical efficacy, setting the stage for future explorations in this exciting interdisciplinary field.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 12 tweets with 409 likes about this paper.