Distill to Delete: Unlearning in Graph Networks with Knowledge Distillation (2309.16173v2)
Abstract: Graph unlearning has emerged as a pivotal method to delete information from a pre-trained graph neural network (GNN). One may delete nodes, a class of nodes, edges, or a class of edges. An unlearning method enables the GNN model to comply with data protection regulations (i.e., the right to be forgotten), adapt to evolving data distributions, and reduce the GPU-hours carbon footprint by avoiding repetitive retraining. Existing partitioning and aggregation-based methods have limitations due to their poor handling of local graph dependencies and additional overhead costs. More recently, GNNDelete offered a model-agnostic approach that alleviates some of these issues. Our work takes a novel approach to address these challenges in graph unlearning through knowledge distillation, as it distills to delete in GNN (D2DGN). It is a model-agnostic distillation framework where the complete graph knowledge is divided and marked for retention and deletion. It performs distillation with response-based soft targets and feature-based node embedding while minimizing KL divergence. The unlearned model effectively removes the influence of deleted graph elements while preserving knowledge about the retained graph elements. D2DGN surpasses the performance of existing methods when evaluated on various real-world graph datasets by up to $43.1\%$ (AUC) in edge and node unlearning tasks. Other notable advantages include better efficiency, better performance in removing target elements, preservation of performance for the retained elements, and zero overhead costs. Notably, our D2DGN surpasses the state-of-the-art GNNDelete in AUC by $2.4\%$, improves membership inference ratio by $+1.3$, requires $10.2\times106$ fewer FLOPs per forward pass and up to $\mathbf{3.2}\times$ faster.
- Machine Unlearning. In 2021 IEEE Symposium on Security and Privacy (SP), 141–159.
- Graph unlearning. In Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 499–513.
- GNNDelete: A General Strategy for Unlearning in Graph Neural Networks.
- Certified Graph Unlearning. In NeurIPS 2022 Workshop: New Frontiers in Graph Learning.
- Efficient Model Updates for Approximate Unlearning of Graph-Structured Data. In International Conference on Learning Representations.
- Can bad teaching induce forgetting? unlearning in deep networks using an incompetent teacher. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, 7210–7217.
- Zero-shot machine unlearning. IEEE Transactions on Information Forensics and Security.
- GraphEditor: An Efficient Graph Representation Learning and Unlearning Approach.
- Efficiently Forgetting What You Have Learned in Graph Representation Learning via Projection. In International Conference on Artificial Intelligence and Statistics, 6674–6703. PMLR.
- Safe: Machine unlearning with shard graphs. arXiv preprint arXiv:2304.13169.
- Magnn: Metapath aggregated graph neural network for heterogeneous graph embedding. In Proceedings of The Web Conference 2020, 2331–2341.
- Goldman, E. 2020. An Introduction to the California Consumer Privacy Act (CCPA). Santa Clara Univ. Legal Studies Research Paper.
- Knowledge distillation: A survey. International Journal of Computer Vision, 129: 1789–1819.
- Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531.
- LGPD Et Al.–Report on the Law of Data Disclosure in Brazil. University Of Passau Irdg Research Paper Series, (22-06).
- Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33: 22118–22133.
- Semi-Supervised Classification with Graph Convolutional Networks. In International Conference on Learning Representations.
- No Matter How You Slice It: Machine Unlearning with SISA Comes at the Expense of Minority Classes. In 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), 622–637.
- Descent-to-delete: Gradient-based methods for machine unlearning. In Algorithmic Learning Theory, 931–962. PMLR.
- A survey of machine unlearning. arXiv preprint arXiv:2209.02299.
- Membership inference attack on graph neural networks. In 2021 Third IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA), 11–20. IEEE.
- Unlearning Graph Classifiers with Limited Data Resources. In Proceedings of the ACM Web Conference 2023, 716–726.
- Distillation as a defense to adversarial perturbations against deep neural networks. In 2016 IEEE symposium on security and privacy (SP), 582–597. IEEE.
- Learn locally, correct globally: A distributed algorithm for training graph neural networks. arXiv preprint arXiv:2111.08202.
- Pitfalls of graph neural network evaluation. arXiv preprint arXiv:1811.05868.
- Deep regression unlearning. In International Conference on Machine Learning, 33921–33939. PMLR.
- Fast Yet Effective Machine Unlearning. IEEE Transactions on Neural Networks and Learning Systems.
- Graph attention networks. stat, 1050(20): 10–48550.
- The eu general data protection regulation (gdpr). A Practical Guide, 1st Ed., Cham: Springer International Publishing.
- Inductive Graph Unlearning. In USENIX Security Symposium.
- Private model compression via knowledge distillation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, 1190–1197.
- Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks. IEEE transactions on pattern analysis and machine intelligence, 44(6): 3048–3068.
- GIF: A General Graph Unlearning Strategy via Influence Function. In Proceedings of the ACM Web Conference 2023, 651–661.
- Certified Edge Unlearning for Graph Neural Networks. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2606–2617.
- A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems, 32(1): 4–24.
- How Powerful are Graph Neural Networks? In International Conference on Learning Representations.
- Revisiting semi-supervised learning with graph embeddings. In International conference on machine learning, 40–48. PMLR.
- Graph neural networks: A review of methods and applications. AI open, 1: 57–81.
- Heterogeneous federated knowledge graph embedding learning and unlearning. In Proceedings of the ACM Web Conference 2023, 2444–2454.