SynHING: Synthetic Heterogeneous Information Network Generation for Graph Learning and Explanation (2401.04133v2)
Abstract: Graph Neural Networks (GNNs) excel in delineating graph structures in diverse domains, including community analysis and recommendation systems. As the interpretation of GNNs becomes increasingly important, the demand for robust baselines and expansive graph datasets is accentuated, particularly in the context of Heterogeneous Information Networks (HIN). Addressing this, we introduce SynHING, a novel framework for Synthetic Heterogeneous Information Network Generation aimed at enhancing graph learning and explanation. SynHING systematically identifies major motifs in a target HIN and employs a bottom-up generation process with intra-cluster and inter-cluster merge modules. This process, supplemented by post-pruning techniques, ensures the synthetic HIN closely mirrors the original graph's structural and statistical properties. Crucially, SynHING provides ground-truth motifs for evaluating GNN explainer models, setting a new standard for explainable, synthetic HIN generation and contributing to the advancement of interpretable machine learning in complex networks.
- Abbe, E. 2017. Community detection and stochastic block models: recent developments. The Journal of Machine Learning Research, 18(1): 6446–6531.
- Statistical mechanics of complex networks. Reviews of modern physics, 74(1): 47.
- Machine learning on graphs: A model and comprehensive taxonomy. Journal of Machine Learning Research, 23(89): 1–64.
- Personalized recommendation system based on collaborative filtering for IoT scenarios. IEEE Transactions on Services Computing, 13(4): 685–695.
- Towards self-explainable graph neural network. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, 302–311.
- Benchmarking graph neural networks.
- TreeXGNN: can gradient-boosted decision trees help boost heterogeneous graph neural networks? In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1–5. IEEE.
- Heterogeneous graph transformer. In Proceedings of the web conference 2020, 2704–2710.
- A survey of explainable graph neural networks: Taxonomy and evaluation metrics. arXiv preprint arXiv:2207.12599.
- Orphicx: A causality-inspired latent variable model for interpreting graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 13729–13738.
- Deep generation of heterogeneous networks. In 2021 IEEE International Conference on Data Mining (ICDM), 379–388. IEEE.
- Parameterized explainer for graph neural network. Advances in neural information processing systems, 33: 19620–19631.
- Are we really making much progress? revisiting, benchmarking and refining heterogeneous graph neural networks. In Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining, 1150–1160.
- Network motifs: simple building blocks of complex networks. Science, 298(5594): 824–827.
- Graphworld: Fake graphs bring real insights for gnns. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 3691–3701.
- Pathfinder discovery networks for neural message passing. In Proceedings of the Web Conference 2021, 2547–2558.
- Overlapping community detection with graph neural networks. arXiv preprint arXiv:1909.12201.
- Estimation and prediction for stochastic blockmodels for graphs with latent block structure. Journal of classification, 14(1): 75–100.
- A deep learning approach to antibiotic discovery. Cell, 180(4): 688–702.
- Graph clustering with graph neural networks. arXiv preprint arXiv:2006.16904.
- Synthetic Graph Generation to Benchmark Graph Learning. arXiv preprint arXiv:2204.01376.
- Heterogeneous Graph Attention Network. In The World Wide Web Conference, WWW ’19, 2022–2032. New York, NY, USA: Association for Computing Machinery. ISBN 9781450366748.
- Gnnexplainer: Generating explanations for graph neural networks. Advances in neural information processing systems, 32.
- On explainability of graph neural networks via subgraph explorations. In International Conference on Machine Learning, 12241–12252. PMLR.
- mg2vec: Learning relationship-preserving heterogeneous graph representations via metagraph embedding. IEEE Transactions on Knowledge and Data Engineering, 34(3): 1317–1329.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.