Distill n' Explain: explaining graph neural networks using simple surrogates
Abstract: Explaining node predictions in graph neural networks (GNNs) often boils down to finding graph substructures that preserve predictions. Finding these structures usually implies back-propagating through the GNN, bonding the complexity (e.g., number of layers) of the GNN to the cost of explaining it. This naturally begs the question: Can we break this bond by explaining a simpler surrogate GNN? To answer the question, we propose Distill n' Explain (DnX). First, DnX learns a surrogate GNN via knowledge distillation. Then, DnX extracts node or edge-level explanations by solving a simple convex program. We also propose FastDnX, a faster version of DnX that leverages the linear decomposition of our surrogate model. Experiments show that DnX and FastDnX often outperform state-of-the-art GNN explainers while being orders of magnitude faster. Additionally, we support our empirical findings with theoretical results linking the quality of the surrogate model (i.e., distillation error) to the faithfulness of explanations.
- Probing GNN Explainers: A Rigorous Theoretical and Empirical Analysis of GNN Explanation Methods. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2022.
- Decomposed knowledge distillation for class-incremental semantic segmentation. In Advances in Neural Information Processing Systems (NeurIPS), 2022.
- Robust counterfactual explanations on graph neural networks. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- A. Barabási and R. Albert. Emergence of scaling in random networks. Science, 286(5439), 1999.
- Graph neural networks with convolutional arma filters. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(7), 2022.
- Simple and deep graph convolutional networks. In International Conference on Machine Learning (ICML), 2020.
- Eta prediction with graph neural networks in google maps. In Conference on Information and Knowledge Management (CIKM), 2021.
- PLOP: Learning without forgetting for continual semantic segmentation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
- When comparing to ground truth is wrong: On evaluating gnn explanation methods. In Conference on Knowledge Discovery & Data Mining (KDD), 2021.
- Degree: Decomposition based explanation for graph neural networks. In International Conference on Learning Representations (ICLR), 2021.
- M. Fey and J. E. Lenssen. Fast graph representation learning with PyTorch Geometric. In Workshop on Representation Learning on Graphs and Manifolds (ICLR), 2019.
- B. Gao and L. Pavel. On the Properties of the Softmax Function with Application in Game Theory and Reinforcement Learning. ArXiv:1704.00805, 2018.
- Neural message passing for quantum chemistry. In International Conference on Machine Learning (ICML), 2017.
- A new model for learning in graph domains. In IEEE International Joint Conference on Neural Networks (IJCNN), 2005.
- W. L. Hamilton. Graph representation learning. Synthesis Lectures on Artificial Intelligence and Machine Learning, 14(3):1–159, 2020.
- Which explanation should i choose? a function approximation perspective to characterizing post hoc explanations. Advances in Neural Information Processing Systems (NeurIPS), 2022.
- Cross-layer distillation with semantic calibration. In AAAI Conference on Artificial Intelligence (AAAI), 2021.
- Distilling the knowledge in a neural network. Arxiv:1503.02531, 2015.
- Combining label propagation and simple models out-performs graph neural networks. In International Conference on Learning Representations (ICLR), 2021.
- Graphlime: Local interpretable model explanations for graph neural networks. IEEE Transactions on Knowledge and Data Engineering, 2022.
- Drug discovery with explainable artificial intelligence. Nature Machine Intelligence, 2(10):573–584, 2020.
- Amalgamating knowledge from heterogeneous graph neural networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
- On representation knowledge distillation for graph neural networks. arXiv:2111.04964, 2021.
- D. Kingma and J. Ba. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), 2015.
- T. N. Kipf and M. Welling. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations (ICLR), 2017.
- Edge weight prediction in weighted signed networks. In International Conference on Data Mining (ICDM), 2016.
- Rev2: Fraudulent user prediction in rating platforms. In International Conference on Web Search and Data Mining (WSDM), 2018.
- Gated graph sequence neural networks. International Conference on Learning Representations (ICLR), 2016.
- Generative causal explanations for graph neural networks. In International Conference on Machine Learning (ICML), 2021.
- Orphicx: A causality-inspired latent variable model for interpreting graph neural networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
- I. Loshchilov and F. Hutter. Decoupled weight decay regularization. In International Conference on Learning Representations (ICLR), 2019.
- Cf-gnnexplainer: Counterfactual explanations for graph neural networks. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2022.
- A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems (NeurIPS), 2017.
- Parameterized explainer for graph neural network. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- Ensemble distribution distillation. In International Conference on Learning Representations (ICLR), 2020.
- Automatic differentiation in pytorch. In Advances in Neural Information Processing Systems (NeurIPS - Workshop), 2017.
- Explainability methods for graph convolutional neural networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- iCaRL: Incremental classifier and representation learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- ”Why should I trust you?”: Explaining the predictions of any classifier. In International Conference on Knowledge Discovery and Data Mining (KDD), 2016.
- SIGN: scalable inception graph neural networks. Workshop on Graph Representation Learning and Beyond (ICML), 2020.
- Scaling ensemble distribution distillation to many classes with proxy targets. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Learning to simulate complex physics with graph networks. In International Conference on Machine Learning (ICML), 2020.
- The graph neural network model. IEEE Transactions on Neural Networks, 20(1), 2009.
- Reliable post hoc explanations: Modeling uncertainty in explainability. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- A deep learning approach to antibiotic discovery. Cell, 180(4), 2020.
- Generalized bayesian posterior expectation distillation for deep neural networks. In Uncertainty in Artificial Intelligence (UAI), 2020.
- R. van de Geijn and M. Myers. Advanced Linear Algebra Foundations to Frontiers. open EdX Publisher, Austin, Texas, 2022.
- M. Vu and M. T. Thai. Pgm-explainer: Probabilistic graphical model explanations for graph neural networks. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- Towards multi-grained explainability for graph neural networks. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Simplifying graph convolutional networks. In International Conference on Machine Learning (ICML), 2019.
- How powerful are graph neural networks? International Conference on Learning Representations (ICLR), 2019.
- Distilling knowledge from graph convolutional networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7074–7083, 2020.
- Graph convolutional neural networks for web-scale recommender systems. In International Conference on Knowledge Discovery & Data Mining (KDD), 2018.
- Gnnexplainer: Generating explanations for graph neural networks. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems (NeurIPS), 2019.
- XGNN: Towards model-level explanations of graph neural networks. In International Conference on Knowledge Discovery & Data Mining (KDD), 2020.
- On explainability of graph neural networks via subgraph explorations. In International Conference on Machine Learning (ICML), 2021.
- Explainability in graph neural networks: A taxonomic survey. IEEE transactions on pattern analysis and machine intelligence, 2022.
- Multi-scale distillation from multiple graph neural networks. In AAAI Conference on Artificial Intelligence (AAAI), 2022a.
- Explaining graph neural networks with structure-aware cooperative games. In Advances in Neural Information Processing Systems (NeurIPS), 2022b.
- Dataset distillation using neural feature regression. In Advances in Neural Information Processing Systems (NeurIPS), 2022.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.