Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Disentangled Condensation for Large-scale Graphs (2401.12231v2)

Published 18 Jan 2024 in cs.SI and cs.LG

Abstract: Graph condensation has emerged as an intriguing technique to save the expensive training costs of Graph Neural Networks (GNNs) by substituting a condensed small graph with the original graph. Despite the promising results achieved, previous methods usually employ an entangled paradigm of redundant parameters (nodes, edges, GNNs), which incurs complex joint optimization during condensation. This paradigm has considerably impeded the scalability of graph condensation, making it challenging to condense extremely large-scale graphs and generate high-fidelity condensed graphs. Therefore, we propose to disentangle the condensation process into a two-stage GNN-free paradigm, independently condensing nodes and generating edges while eliminating the need to optimize GNNs at the same time. The node condensation module avoids the complexity of GNNs by focusing on node feature alignment with anchors of the original graph, while the edge translation module constructs the edges of the condensed nodes by transferring the original structure knowledge with neighborhood anchors. This simple yet effective approach achieves at least 10 times faster than state-of-the-art methods with comparable accuracy on medium-scale graphs. Moreover, the proposed DisCo can successfully scale up to the Ogbn-papers100M graph with flexible reduction rates. Extensive downstream tasks and ablation study on five common datasets further demonstrate the effectiveness of the proposed DisCo framework. The source code will be made publicly available.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Shaked Brody, Uri Alon and Eran Yahav “How Attentive are Graph Attention Networks?” In International Conference on Learning Representations, 2021
  2. “Dataset distillation by matching training trajectories” In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4750–4759
  3. “Graph neural network for fraud detection via spatial-temporal attention” In IEEE Transactions on Knowledge and Data Engineering 34.8, 2020, pp. 3800–3813
  4. “One trillion edges: Graph processing at facebook-scale” In Proceedings of the VLDB Endowment 8.12, 2015, pp. 1804–1815
  5. “Remember the past: Distilling datasets into addressable memories for neural networks” In arXiv preprint arXiv:2206.02916, 2022
  6. Tian Dong, Bo Zhao and Lingjuan Lyu “Privacy for free: How does dataset condensation help privacy?” In International Conference on Machine Learning, 2022, pp. 5378–5396
  7. “Graph neural networks for social recommendation” In International World Wide Web Conference, 2019, pp. 417–426
  8. “Graph Condensation for Inductive Node Representation Learning” In arXiv preprint arXiv:2307.15967, 2023
  9. “Graph neural architecture search” In International Joint Conference on Artificial Intelligence, 2021
  10. William L. Hamilton, Zhitao Ying and Jure Leskovec “Inductive Representation Learning on Large Graphs” In Advances in Neural Information Processing Systems, 2017
  11. “Open graph benchmark: Datasets for machine learning on graphs” In Conference on Neural Information Processing Systems 33, 2020, pp. 22118–22133
  12. “Condensing graphs via one-step gradient matching” In ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2022, pp. 720–730
  13. “Graph condensation for graph neural networks” In arXiv preprint arXiv:2110.07580, 2021
  14. “Amalgamating knowledge from heterogeneous graph neural networks” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 15709–15718
  15. Thomas N. Kipf and Max Welling “Semi-Supervised Classification with Graph Convolutional Networks” In International Conference on Learning Representations, 2017
  16. “Graph condensation via receptive field distribution matching” In arXiv preprint arXiv:2206.13697, 2022
  17. “Slimmable dataset condensation” In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 3759–3768
  18. Andreas Loukas “Graph Reduction with Spectral and Cut Guarantees.” In J. Mach. Learn. Res. 20.116, 2019, pp. 1–42
  19. “Spectrally Approximating Large Graphs with Smaller Graphs” In International Conference on Machine Learning, Proceedings of Machine Learning Research, 2018
  20. Alexander Mordvintsev, Christopher Olah and Mike Tyka “Inceptionism: Going deeper into neural networks”, 2015
  21. Timothy Nguyen, Zhourong Chen and Jaehoon Lee “Dataset meta-learning from kernel ridge-regression” In arXiv preprint arXiv:2011.00050, 2020
  22. “Dataset distillation with infinitely wide convolutional networks”, 2021, pp. 5186–5198
  23. David Peleg and Alejandro A Schäffer “Graph spanners” In Journal of Graph Theory 13.1, 1989, pp. 99–116
  24. “Graph neural networks for materials science and chemistry” In Communications Materials 3.1, 2022, pp. 93
  25. “Active learning for convolutional neural networks: A core-set approach” In arXiv preprint arXiv:1708.00489, 2017
  26. Daniel A Spielman and Shang-Hua Teng “Spectral sparsification of graphs” In SIAM Journal on Computing 40.4, 2011, pp. 981–1025
  27. “Generative teaching networks: Accelerating neural architecture search by learning to generate synthetic training data” In International Conference on Machine Learning, 2020, pp. 9206–9216
  28. “Cafe: Learning to condense dataset by aligning features” In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 12196–12205
  29. “Dataset distillation” In arXiv preprint arXiv:1811.10959, 2018
  30. Max Welling “Herding dynamical weights to learn” In International Conference on Machine Learning, 2009, pp. 1121–1128
  31. “Simplifying graph convolutional networks” In International Conference on Machine Learning, 2019, pp. 6861–6871
  32. “Rumor detection based on propagation graph neural network with attention mechanism” In Expert systems with applications 158, 2020, pp. 113595
  33. “A comprehensive survey on graph neural networks” In IEEE Transactions on Neural Networks and Learning Systems 32.1, 2020, pp. 4–24
  34. “How Powerful are Graph Neural Networks?” In International Conference on Learning Representations, 2018
  35. “Representation learning on graphs with jumping knowledge networks” In International Conference on Machine Learning, 2018, pp. 5453–5462
  36. “Do transformers really perform badly for graph representation?” In Advances in Neural Information Processing Systems 34, 2021, pp. 28877–28888
  37. “Graph convolutional neural networks for web-scale recommender systems” In ACM SIGKDD international Conference on Knowledge Discovery and Data Mining, 2018, pp. 974–983
  38. “Hierarchical graph representation learning with differentiable pooling” In Advances in neural information processing systems 31, 2018
  39. “Graphsaint: Graph sampling based inductive learning method” In arXiv preprint arXiv:1907.04931, 2019
  40. “Link prediction based on graph neural networks” In Conference on Neural Information Processing Systems 31, 2018
  41. “Dataset condensation with differentiable siamese augmentation” In International Conference on Machine Learning, 2021, pp. 12674–12685
  42. “Dataset condensation with distribution matching” In IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 6514–6523
  43. Bo Zhao, Konda Reddy Mopuri and Hakan Bilen “Dataset Condensation with Gradient Matching” In International Conference on Learning Representations, 2021
  44. “Structure-free Graph Condensation: From Large-scale Graphs to Condensed Graph-free Data” In arXiv preprint arXiv:2306.02664, 2023
  45. “TGL: a general framework for temporal GNN training on billion-scale graphs” In Proceedings of the VLDB Endowment 15.8, 2022, pp. 1572–1580
  46. “Graph neural networks: A review of methods and applications” In AI Open 1, 2020, pp. 57–81
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com