Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Distance-Restricted Folklore Weisfeiler-Leman GNNs with Provable Cycle Counting Power (2309.04941v3)

Published 10 Sep 2023 in cs.LG

Abstract: The ability of graph neural networks (GNNs) to count certain graph substructures, especially cycles, is important for the success of GNNs on a wide range of tasks. It has been recently used as a popular metric for evaluating the expressive power of GNNs. Many of the proposed GNN models with provable cycle counting power are based on subgraph GNNs, i.e., extracting a bag of subgraphs from the input graph, generating representations for each subgraph, and using them to augment the representation of the input graph. However, those methods require heavy preprocessing, and suffer from high time and memory costs. In this paper, we overcome the aforementioned limitations of subgraph GNNs by proposing a novel class of GNNs -- $d$-Distance-Restricted FWL(2) GNNs, or $d$-DRFWL(2) GNNs. $d$-DRFWL(2) GNNs use node pairs whose mutual distances are at most $d$ as the units for message passing to balance the expressive power and complexity. By performing message passing among distance-restricted node pairs in the original graph, $d$-DRFWL(2) GNNs avoid the expensive subgraph extraction operations in subgraph GNNs, making both the time and space complexity lower. We theoretically show that the discriminative power of $d$-DRFWL(2) GNNs strictly increases as $d$ increases. More importantly, $d$-DRFWL(2) GNNs have provably strong cycle counting power even with $d=2$: they can count all 3, 4, 5, 6-cycles. Since 6-cycles (e.g., benzene rings) are ubiquitous in organic molecules, being able to detect and count them is crucial for achieving robust and generalizable performance on molecular tasks. Experiments on both synthetic datasets and molecular datasets verify our theory. To the best of our knowledge, our model is the most efficient GNN model to date (both theoretically and empirically) that can count up to 6-cycles.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. The surprising power of graph neural networks with random node initialization. arXiv preprint arXiv:2010.01179, 2020.
  2. Finding and counting given length cycles. Algorithmica, 17(3):209–223, 1997.
  3. Cormorant: Covariant molecular neural networks. Advances in neural information processing systems, 32, 2019.
  4. On weisfeiler-leman invariance: Subgraph counts and related graph properties. Journal of Computer and System Sciences, 113:42–59, 2020a. ISSN 0022-0000. doi: https://doi.org/10.1016/j.jcss.2020.04.003. URL https://www.sciencedirect.com/science/article/pii/S0022000020300386.
  5. On weisfeiler-leman invariance: Subgraph counts and related graph properties. Journal of Computer and System Sciences, 113:42–59, 2020b.
  6. Breaking the limits of message passing graph neural networks. In International Conference on Machine Learning, pages 599–608. PMLR, 2021.
  7. Directional graph networks. In International Conference on Machine Learning, pages 748–758. PMLR, 2021.
  8. Equivariant subgraph aggregation networks. In International Conference on Learning Representations, 2022.
  9. Weisfeiler and lehman go cellular: Cw networks. Advances in Neural Information Processing Systems, 34:2625–2640, 2021.
  10. Improving graph neural network expressivity via subgraph isomorphism counting. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):657–668, 2022.
  11. Residual gated graph convnets. arXiv preprint arXiv:1711.07553, 2017.
  12. An optimal lower bound on the number of variables for graph identification. Combinatorica, 12(4):389–410, 1992.
  13. Simple and deep graph convolutional networks. In International conference on machine learning, pages 1725–1735. PMLR, 2020a.
  14. On the equivalence between graph isomorphism testing and function approximation with gnns. Advances in neural information processing systems, 32, 2019.
  15. Can graph neural networks count substructures? Advances in neural information processing systems, 33:10383–10395, 2020b.
  16. Principal neighbourhood aggregation for graph nets. Advances in Neural Information Processing Systems, 33:13260–13271, 2020.
  17. Reconstruction for powerful graph representations. Advances in Neural Information Processing Systems, 34:1713–1726, 2021.
  18. Automated approaches for classifying structures. Technical report, MINNESOTA UNIV MINNEAPOLIS DEPT OF COMPUTER SCIENCE, 2002.
  19. A generalization of transformer networks to graphs. arXiv preprint arXiv:2012.09699, 2020.
  20. Benchmarking graph neural networks. 2020.
  21. Long range graph benchmark. Advances in Neural Information Processing Systems, 35:22326–22340, 2022.
  22. Fast graph representation learning with pytorch geometric. arXiv preprint arXiv:1903.02428, 2019.
  23. Hierarchical inter-message passing for learning on molecular graphs. arXiv preprint arXiv:2006.12179, 2020.
  24. Understanding and extending subgraph gnns by rethinking their symmetries. In Advances in Neural Information Processing Systems, 2022.
  25. Martin Fürer. On the combinatorial power of the weisfeiler-lehman algorithm. In International Conference on Algorithms and Complexity, pages 260–271. Springer, 2017.
  26. Directional message passing for molecular graphs. arXiv preprint arXiv:2003.03123, 2020.
  27. Neural message passing for quantum chemistry. In International conference on machine learning, pages 1263–1272. PMLR, 2017.
  28. Pebble games and linear equations. The Journal of Symbolic Logic, 80(3):797–844, 2015.
  29. Intrinsic-extrinsic convolution and pooling for learning on 3d protein structures. arXiv preprint arXiv:2007.06252, 2020.
  30. Strategies for pre-training graph neural networks. arXiv preprint arXiv:1905.12265, 2019.
  31. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118–22133, 2020.
  32. A short tutorial on the weisfeiler-lehman test and its variants. In ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 8533–8537, 2021. doi: 10.1109/ICASSP39728.2021.9413523.
  33. Boosting the cycle counting power of graph neural networks with i22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT-gnns. 2023.
  34. Finding frequent subgraphs in longitudinal social network data using a weighted graph mining approach. In Advanced Data Mining and Applications: 6th International Conference, ADMA 2010, Chongqing, China, November 19-21, 2010, Proceedings, Part I 6, pages 405–416. Springer, 2010.
  35. Junction tree variational autoencoder for molecular graph generation. In International conference on machine learning, pages 2323–2332. PMLR, 2018.
  36. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
  37. An efficient algorithm for detecting frequent subgraphs in biological networks. Bioinformatics, 20(suppl_1):i200–i207, 2004.
  38. Rethinking graph transformers with spectral attention. Advances in Neural Information Processing Systems, 34:21618–21629, 2021.
  39. Distance encoding: Design provably more powerful neural networks for graph representation learning. Advances in Neural Information Processing Systems, 33:4465–4478, 2020.
  40. Spherical message passing for 3d graph networks. arXiv preprint arXiv:2102.05013, 2021.
  41. Invariant and equivariant graph networks. arXiv preprint arXiv:1812.09902, 2018.
  42. Provably powerful graph networks. Advances in neural information processing systems, 32, 2019a.
  43. On the universality of invariant networks. In International conference on machine learning, pages 4363–4371. PMLR, 2019b.
  44. Weisfeiler and leman go neural: Higher-order graph neural networks. In Proceedings of the AAAI conference on artificial intelligence, pages 4602–4609, 2019.
  45. Weisfeiler and leman go sparse: Towards scalable higher-order graph embeddings. Advances in Neural Information Processing Systems, 33:21824–21840, 2020.
  46. Ordered subgraph aggregation networks. In Advances in Neural Information Processing Systems, 2022.
  47. Orbnet: Deep learning for quantum chemistry using symmetry-adapted atomic-orbital features. The Journal of chemical physics, 153(12):124111, 2020.
  48. Counting substructures with higher-order graph neural networks: Possibility and impossibility results. arXiv preprint arXiv:2012.03174, 2020.
  49. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  50. Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.
  51. Towards better evaluation of gnn expressiveness with brec dataset. arXiv preprint arXiv:2304.07702, 2023.
  52. The reduction of a graph to canonical form and the algebra which appears therein. NTI, Series, 2(9):12–16, 1968.
  53. Moleculenet: a benchmark for molecular machine learning. Chemical science, 9(2):513–530, 2018.
  54. A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems, 32(1):4–24, 2020.
  55. How powerful are graph neural networks? In International Conference on Learning Representations, 2018.
  56. Efficiently counting substructures by subgraph gnns without running gnn on subgraphs. arXiv preprint arXiv:2303.10576, 2023.
  57. Identity-aware graph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 10737–10745, 2021.
  58. A complete expressiveness hierarchy for subgraph gnns via subgraph weisfeiler-lehman tests. arXiv preprint arXiv:2302.07090, 2023.
  59. Nested graph neural networks. Advances in Neural Information Processing Systems, 34:15734–15747, 2021.
  60. From stars to subgraphs: Uplifting any gnn with local structure awareness. In International Conference on Learning Representations, 2022.
  61. Graph neural networks: A review of methods and applications. AI open, 1:57–81, 2020.
Citations (5)

Summary

We haven't generated a summary for this paper yet.