Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Remaining-data-free Machine Unlearning by Suppressing Sample Contribution (2402.15109v3)

Published 23 Feb 2024 in cs.LG

Abstract: Machine unlearning (MU) is to forget data from a well-trained model, which is practically important due to the ``right to be forgotten''. The unlearned model should approach the retrained model, where the forgetting data are not involved in the training process and hence do not contribute to the retrained model. Considering the forgetting data's absence during retraining, we think unlearning should withdraw their contribution from the pre-trained model. The challenge is that when tracing the learning process is impractical, how to quantify and detach sample's contribution to the dynamic learning process using only the pre-trained model. We first theoretically discover that sample's contribution during the process will reflect in the learned model's sensitivity to it. We then practically design a novel method, namely MU-Mis (Machine Unlearning by Minimizing input sensitivity), to suppress the contribution of the forgetting data. Experimental results demonstrate that MU-Mis can unlearn effectively and efficiently without utilizing the remaining data. It is the first time that a remaining-data-free method can outperform state-of-the-art (SoTA) unlearning methods that utilize the remaining data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Where is the information in a deep neural network? arXiv preprint arXiv:1905.12213, 2019.
  2. Influence functions in deep learning are fragile. In International Conference on Learning Representations, 2020.
  3. Burak. Pinterest face recognition dataset. www.kaggle.com/datasets/hereisburak/pins-facerecognition, 2020.
  4. Towards making systems forget with machine unlearning. In IEEE Symposium on Security and Privacy, pp.  463–480. IEEE, 2015.
  5. Support Vector Machines. Springer, 2008.
  6. Can bad teaching induce forgetting? unlearning in deep networks using an incompetent teacher. In the AAAI Conference on Artificial Intelligence, pp.  7210–7217, 2023.
  7. Support-vector networks. Machine Learning, 20:273–297, 1995.
  8. Salun: Empowering machine unlearning via gradient-based weight saliency in both image classification and generation. arXiv preprint arXiv:2310.12508, 2023.
  9. Feldman, V. Does learning require memorization? a short tale about a long tail. In the Annual ACM SIGACT Symposium on Theory of Computing, pp.  954–959, 2020.
  10. What neural networks memorize and why: Discovering the long tail via influence estimation. Advances in Neural Information Processing Systems, 33:2881–2891, 2020.
  11. Emergent properties of the local geometry of neural loss landscapes. arXiv preprint arXiv:1910.05929, 2019.
  12. Fast machine unlearning without retraining through selective synaptic dampening. arXiv preprint arXiv:2308.07707, 2023.
  13. Eternal sunshine of the spotless net: Selective forgetting in deep networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  9304–9312, 2020a.
  14. Forgetting outside the box: Scrubbing deep networks of information accessible from input-output observations. In European Conference on Computer Vision, pp.  383–398, 2020b.
  15. Amnesiac machine learning. In the AAAI Conference on Artificial Intelligence, pp.  11516–11524, 2021.
  16. Certified data removal from machine learning models. In Proceedings of the 37th International Conference on Machine Learning, pp.  3832–3842, 2020.
  17. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition, pp.  770–778, 2016.
  18. Learn to unlearn for deep neural networks: Minimizing unlearning interference with gradient projection. In IEEE/CVF Winter Conference on Applications of Computer Vision, pp.  4819–4828, 2024.
  19. Understanding black-box predictions via influence functions. In International Conference on Machine Learning, pp.  1885–1894. PMLR, 2017.
  20. Deep unlearning via randomized conditionally independent hessians. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  10422–10431, 2022.
  21. Descent-to-delete: Gradient-based methods for machine unlearning. In Algorithmic Learning Theory, pp.  931–962. PMLR, 2021.
  22. Papyan, V. Traces of class/cross-class structure pervade deep learning spectra. Journal of Machine Learning Research, 21(1):10197–10260, 2020.
  23. SSSE: Efficiently erasing samples from trained machine learning models. In NeurIPS 2021 Workshop Privacy in Machine Learning, 2021.
  24. Regulation, G. D. P. General data protection regulation (GDPR). Intersoft Consulting, Accessed in October, 24(1), 2018.
  25. Grad-cam: Visual explanations from deep networks via gradient-based localization. In IEEE International Conference on Computer Vision, pp.  618–626, 2017.
  26. Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy (SP), pp.  3–18. IEEE, 2017.
  27. Smoothgrad: removing noise by adding noise. arXiv preprint arXiv:1706.03825, 2017.
  28. Algorithms that approximate data removal: New results and limitations. Advances in Neural Information Processing Systems, pp.  18892–18903, 2022.
  29. Fast yet effective machine unlearning. IEEE Transactions on Neural Networks and Learning Systems, 2023.
  30. Puma: Performance unchanged model augmentation for training data removal. In the AAAI Conference on Artificial Intelligence, volume 36, pp.  8675–8682, 2022.
  31. Deltagrad: Rapid retraining of machine learning models. In International Conference on Machine Learning, pp.  10355–10366. PMLR, 2020.
  32. Machine unlearning: Solutions and challenges. arXiv preprint arXiv:2308.07061, 2023.
  33. Counterfactual memorization in neural language models. arXiv preprint arXiv:2112.12938, 2021.
  34. Deep leakage from gradients. Advances in Neural Information Processing Systems, 32, 2019.
Citations (2)

Summary

We haven't generated a summary for this paper yet.