Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mirage: Model-Agnostic Graph Distillation for Graph Classification (2310.09486v4)

Published 14 Oct 2023 in cs.LG and cs.AI

Abstract: GNNs, like other deep learning models, are data and computation hungry. There is a pressing need to scale training of GNNs on large datasets to enable their usage on low-resource environments. Graph distillation is an effort in that direction with the aim to construct a smaller synthetic training set from the original training data without significantly compromising model performance. While initial efforts are promising, this work is motivated by two key observations: (1) Existing graph distillation algorithms themselves rely on training with the full dataset, which undermines the very premise of graph distillation. (2) The distillation process is specific to the target GNN architecture and hyper-parameters and thus not robust to changes in the modeling pipeline. We circumvent these limitations by designing a distillation algorithm called Mirage for graph classification. Mirage is built on the insight that a message-passing GNN decomposes the input graph into a multiset of computation trees. Furthermore, the frequency distribution of computation trees is often skewed in nature, enabling us to condense this data into a concise distilled summary. By compressing the computation data itself, as opposed to emulating gradient flows on the original training set-a prevalent approach to date-Mirage transforms into an unsupervised and architecture-agnostic distillation algorithm. Extensive benchmarking on real-world datasets underscores Mirage's superiority, showcasing enhanced generalization accuracy, data compression, and distillation efficiency when compared to state-of-the-art baselines.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Learning articulated rigid body dynamics with lagrangian graph neural network. Advances in Neural Information Processing Systems, 35:29789–29800, 2022.
  2. Enhancing the inductive biases of graph neural ode for modeling dynamical systems. ICLR, 2023.
  3. Tree isomorphism algorithms: Speed vs. clarity. Mathematics Magazine, 64(4):252–261, 1991. ISSN 0025570X, 19300980. URL http://www.jstor.org/stable/2690833.
  4. Learning and maximizing influence in social networks under capacity constraints. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, pp.  733–741, 2023.
  5. Canonical forms for labelled trees and their applications in frequent subtree mining. Knowledge and Information Systems, 8(2):203–234, Aug 2005. ISSN 0219-3116. doi: 10.1007/s10115-004-0180-7. URL https://doi.org/10.1007/s10115-004-0180-7.
  6. Graphgen: a scalable approach to domain-agnostic labeled graph generation. In Proceedings of The Web Conference 2020, pp.  1253–1263, 2020.
  7. FRIGATE: Frugal spatio-temporal forecasting on road networks. In 29th SIGKDD Conference on Knowledge Discovery and Data Mining, 2023. URL https://openreview.net/forum?id=2cTw2M47L1.
  8. Tigger: Scalable generative modelling for temporal interaction graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp.  6819–6828, 2022.
  9. Inductive representation learning on large graphs. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, pp.  1025–1035, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
  10. Mining frequent patterns without candidate generation: A frequent-pattern tree approach. Data Mining and Knowledge Discovery, 8(1):53–87, January 2004.
  11. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118–22133, 2020.
  12. Neuromlr: Robust & reliable route recommendation on road networks. Advances in Neural Information Processing Systems, 34:22070–22082, 2021.
  13. Graph condensation for graph neural networks. In International Conference on Learning Representations, 2021.
  14. Condensing graphs via one-step gradient matching. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp.  720–730, 2022.
  15. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
  16. Spatial transition learning on road networks with deep probabilistic models. 2020 IEEE 36th International Conference on Data Engineering (ICDE), pp.  349–360, 2020. URL https://api.semanticscholar.org/CorpusID:218906673.
  17. Gcomb: Learning budget-constrained combinatorial algorithms over billion-sized graphs. Advances in Neural Information Processing Systems, 33:20000–20011, 2020.
  18. Generative modeling of labeled graphs under data scarcity. In The Second Learning on Graphs Conference, 2023. URL https://openreview.net/forum?id=Hy9K2WiVwW.
  19. Tudataset: A collection of benchmark datasets for learning with graphs. arXiv preprint arXiv:2007.08663, 2020.
  20. Recipe for a General, Powerful, Scalable Graph Transformer. Advances in Neural Information Processing Systems, 35, 2022.
  21. Learning to simulate complex physics with graph networks. In Hal Daumé III and Aarti Singh (eds.), Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pp.  8459–8468. PMLR, 13–18 Jul 2020.
  22. E(n) equivariant graph neural networks. In ICML, 2022.
  23. Weisfeiler-lehman graph kernels. J. Mach. Learn. Res., 12(null):2539–2561, nov 2011. ISSN 1532-4435.
  24. Exphormer: Sparse transformers for graphs. In ICML, 2023.
  25. Understanding over-squashing and bottlenecks on graphs via curvature. In ICLR, 2022.
  26. Gabriel Valiente. Tree Isomorphism, pp.  151–251. Springer Berlin Heidelberg, Berlin, Heidelberg, 2002. ISBN 978-3-662-04921-1. doi: 10.1007/978-3-662-04921-1˙4. URL https://doi.org/10.1007/978-3-662-04921-1˙4.
  27. Graph attention networks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rJXMpikCZ.
  28. Digress: Discrete denoising diffusion for graph generation. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=UaAD-Nu86WX.
  29. Influence maximization in multi-relational social networks. In Proceedings of the 30th ACM international conference on information & knowledge management, pp.  4193–4202, 2021.
  30. Max Welling. Herding dynamical weights to learn. In Proceedings of the 26th Annual International Conference on Machine Learning, pp.  1121–1128, 2009.
  31. Modeling trajectories with recurrent neural networks. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, IJCAI’17, pp.  3083–3090, 2017.
  32. How powerful are graph neural networks? In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=ryGs6iA5Km.
  33. Kernel ridge regression-based graph dataset distillation. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD ’23, pp.  2850–2861, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9798400701030. doi: 10.1145/3580305.3599398. URL https://doi.org/10.1145/3580305.3599398.
  34. Does graph distillation see like vision dataset counterpart? In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=VqIWgUVsXc.
  35. Like like alike: joint friendship and interest propagation in social networks. In Proceedings of the 20th international conference on World wide web, pp.  537–546, 2011.
  36. Do transformers really perform badly for graph representation? In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=OeWooOxFwDa.
  37. Graphrnn: Generating realistic graphs with deep auto-regressive models. In Jennifer G. Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pp.  5694–5703. PMLR, 2018. URL http://proceedings.mlr.press/v80/you18a.html.
  38. Dataset condensation with gradient matching. In ICLR, 2021.
Citations (7)

Summary

We haven't generated a summary for this paper yet.