Langevin Unlearning: A New Perspective of Noisy Gradient Descent for Machine Unlearning (2401.10371v5)
Abstract: Machine unlearning has raised significant interest with the adoption of laws ensuring the ``right to be forgotten''. Researchers have provided a probabilistic notion of approximate unlearning under a similar definition of Differential Privacy (DP), where privacy is defined as statistical indistinguishability to retraining from scratch. We propose Langevin unlearning, an unlearning framework based on noisy gradient descent with privacy guarantees for approximate unlearning problems. Langevin unlearning unifies the DP learning process and the privacy-certified unlearning process with many algorithmic benefits. These include approximate certified unlearning for non-convex problems, complexity saving compared to retraining, sequential and batch unlearning for multiple unlearning requests.
- Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, pp. 308–318, 2016.
- Privacy of noisy stochastic gradient descent: More iterations without more privacy loss. Advances in Neural Information Processing Systems, 35:3788–3800, 2022a.
- Resolving the mixing time of the langevin algorithm to its stationary distribution for log-concave sampling. arXiv preprint arXiv:2210.08448, 2022b.
- Machine unlearning. In 2021 IEEE Symposium on Security and Privacy (SP), pp. 141–159. IEEE, 2021.
- Towards making systems forget with machine unlearning. In 2015 IEEE symposium on security and privacy, pp. 463–480. IEEE, 2015.
- The secret sharer: Evaluating and testing unintended memorization in neural networks. In 28th USENIX Security Symposium (USENIX Security 19), pp. 267–284, 2019.
- Dimension-free log-sobolev inequalities for mixture distributions. Journal of Functional Analysis, 281(11):109236, 2021.
- Efficient model updates for approximate unlearning of graph-structured data. In The Eleventh International Conference on Learning Representations, 2022.
- Forget unlearning: Towards true data-deletion in machine learning. In International Conference on Machine Learning, pp. 6028–6073. PMLR, 2023.
- Differential privacy dynamics of langevin diffusion and noisy gradient descent. Advances in Neural Information Processing Systems, 34:14771–14781, 2021.
- Sparse regression learning by aggregation and langevin monte-carlo. Journal of Computer and System Sciences, 78(5):1423–1443, 2012.
- Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142, 2012.
- Nonasymptotic convergence analysis for the unadjusted langevin algorithm. 2017.
- Calibrating noise to sensitivity in private data analysis. In Theory of Cryptography: Third Theory of Cryptography Conference, TCC 2006, New York, NY, USA, March 4-7, 2006. Proceedings 3, pp. 265–284. Springer, 2006.
- Convergence of langevin monte carlo in chi-squared and rényi divergence. In International Conference on Artificial Intelligence and Statistics, pp. 8151–8175. PMLR, 2022.
- Privacy amplification by iteration. In 2018 IEEE 59th Annual Symposium on Foundations of Computer Science (FOCS), pp. 521–532. IEEE, 2018.
- Bayesian inference forgetting. arXiv preprint arXiv:2101.06417, 2021.
- Faster differentially private samplers via rényi divergence analysis of discretized langevin mcmc. Advances in Neural Information Processing Systems, 33:7222–7233, 2020.
- Gronwall, T. H. Note on the derivatives with respect to a parameter of the solutions of a system of differential equations. Annals of Mathematics, pp. 292–296, 1919.
- Gross, L. Logarithmic sobolev inequalities. American Journal of Mathematics, 97(4):1061–1083, 1975.
- Certified data removal from machine learning models. In International Conference on Machine Learning, pp. 3832–3842. PMLR, 2020.
- Adaptive machine unlearning. Advances in Neural Information Processing Systems, 34:16319–16330, 2021.
- Train faster, generalize better: Stability of stochastic gradient descent. In International conference on machine learning, pp. 1225–1234. PMLR, 2016.
- Hastings, W. K. Monte carlo sampling methods using markov chains and their applications. 1970.
- Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- The composition theorem for differential privacy. In International conference on machine learning, pp. 1376–1385. PMLR, 2015.
- Practical and private (deep) learning without sampling or shuffling. In International Conference on Machine Learning, pp. 5213–5225. PMLR, 2021.
- Krizhevsky, A. et al. Learning multiple layers of features from tiny images. 2009.
- Lamperski, A. Projected stochastic gradient langevin algorithms for constrained sampling and non-convex learning. In Conference on Learning Theory, pp. 2891–2937. PMLR, 2021.
- Sampling can be faster than optimization. Proceedings of the National Academy of Sciences, 116(42):20881–20885, 2019.
- maintainers, T. and contributors. Torchvision: Pytorch’s computer vision library. https://github.com/pytorch/vision, 2016.
- Markov chains and stochastic stability. Springer Science & Business Media, 2012.
- Mironov, I. Rényi differential privacy. In 2017 IEEE 30th computer security foundations symposium (CSF), pp. 263–275. IEEE, 2017.
- Towards a complete analysis of langevin monte carlo: Beyond poincar\\\backslash\’e inequality. arXiv preprint arXiv:2303.03589, 2023.
- Neal, R. M. et al. Mcmc using hamiltonian dynamics. Handbook of markov chain monte carlo, 2(11):2, 2011.
- Descent-to-delete: Gradient-based methods for machine unlearning. In Algorithmic Learning Theory, pp. 931–962. PMLR, 2021.
- Variational bayesian unlearning. Advances in Neural Information Processing Systems, 33:16025–16036, 2020.
- Markov chain monte carlo-based machine unlearning: Unlearning what needs to be forgotten. In Proceedings of the 2022 ACM on Asia Conference on Computer and Communications Security, pp. 351–363, 2022.
- Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pp. 8024–8035. Curran Associates, Inc., 2019.
- Challenges and pitfalls of bayesian unlearning. arXiv preprint arXiv:2207.03227, 2022.
- Differential privacy guarantees for stochastic gradient langevin dynamics. arXiv preprint arXiv:2201.11980, 2022.
- Remember what you want to forget: Algorithms for machine unlearning. Advances in Neural Information Processing Systems, 34:18075–18086, 2021.
- Sinho Chewi. Log-Concave Sampling. https://chewisinho.github.io/main.pdf, 2023. Online; accessed September 29, 2023.
- From adaptive query release to machine unlearning. In International Conference on Machine Learning, pp. 34642–34667. PMLR, 2023.
- Machine unlearning via algorithmic stability. In Conference on Learning Theory, pp. 4126–4142. PMLR, 2021.
- Rapid convergence of the unadjusted langevin algorithm: Isoperimetry suffices. Advances in neural information processing systems, 32, 2019.
- Differentially private learning needs hidden state (or much faster convergence). Advances in Neural Information Processing Systems, 35:703–715, 2022.