Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 69 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 37 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 119 tok/s Pro
Kimi K2 218 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4.5 33 tok/s Pro
2000 character limit reached

Clipped SGD Algorithms for Performative Prediction: Tight Bounds for Clipping Bias and Remedies (2404.10995v2)

Published 17 Apr 2024 in math.OC, cs.CR, and cs.LG

Abstract: This paper studies the convergence of clipped stochastic gradient descent (SGD) algorithms with decision-dependent data distribution. Our setting is motivated by privacy preserving optimization algorithms that interact with performative data where the prediction models can influence future outcomes. This challenging setting involves the non-smooth clipping operator and non-gradient dynamics due to distribution shifts. We make two contributions in pursuit for a performative stable solution using clipped SGD algorithms. First, we characterize the clipping bias with projected clipped SGD (PCSGD) algorithm which is caused by the clipping operator that prevents PCSGD from reaching a stable solution. When the loss function is strongly convex, we quantify the lower and upper bounds for this clipping bias and demonstrate a bias amplification phenomenon with the sensitivity of data distribution. When the loss function is non-convex, we bound the magnitude of stationarity bias. Second, we propose remedies to mitigate the bias either by utilizing an optimal step size design for PCSGD, or to apply the recent DiceSGD algorithm [Zhang et al., 2024]. Our analysis is also extended to show that the latter algorithm is free from clipping bias in the performative setting. Numerical experiments verify our findings.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, pages 308–318, 2016.
  2. Private empirical risk minimization: Efficient algorithms and tight error bounds. In 2014 IEEE 55th annual symposium on foundations of computer science, pages 464–473. IEEE, 2014.
  3. Performative prediction in a stateful world. In International Conference on Artificial Intelligence and Statistics, pages 6045–6061. PMLR, 2022.
  4. Differentially private empirical risk minimization. Journal of Machine Learning Research, 12(3), 2011.
  5. Understanding gradient clipping in private sgd: A geometric perspective. Advances in Neural Information Processing Systems, 33:13773–13782, 2020.
  6. Stochastic optimization with decision-dependent distributions. Mathematics of Operations Research, 48(2):954–998, 2023.
  7. Cynthia Dwork. Differential privacy. In Automata, Languages and Programming: 33rd International Colloquium, ICALP 2006, Venice, Italy, July 10-14, 2006, Proceedings, Part II 33, pages 1–12. Springer, 2006.
  8. The algorithmic foundations of differential privacy. Foundations and Trends® in Theoretical Computer Science, 9(3–4):211–407, 2014.
  9. Private stochastic convex optimization: optimal rates in linear time. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 439–449, 2020.
  10. Differential privacy for stackelberg games. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20), pages 3452–3458, 07 2020.
  11. The optimal noise-adding mechanism in differential privacy. IEEE Transactions on Information Theory, 62(2):925–951, 2015.
  12. Universally utility-maximizing privacy mechanisms. In Proceedings of the forty-first annual ACM symposium on Theory of computing, pages 351–360, 2009.
  13. Stochastic optimization with heavy-tailed noise via accelerated gradient clipping. Advances in Neural Information Processing Systems, 33:15042–15053, 2020.
  14. Beyond convexity: Stochastic quasi-convex optimization. Advances in neural information processing systems, 28, 2015.
  15. How to learn when data reacts to your model: performative gradient descent. In International Conference on Machine Learning, pages 4641–4650. PMLR, 2021.
  16. Kaggle. Give me some credit, Improve on the state of the art in credit scoring by predicting the probability that somebody will experience financial distress in the next two years., 2011. URL https://www.kaggle.com/c/GiveMeSomeCredit.
  17. Clip21: Error feedback for gradient clipping. arXiv preprint arXiv:2305.18929, 2023.
  18. Revisiting gradient clipping: Stochastic bias and tight convergence guarantees. arXiv preprint arXiv:2305.01588, 2023.
  19. State dependent performative prediction with stochastic approximation. In International Conference on Artificial Intelligence and Statistics, pages 3164–3186. PMLR, 2022.
  20. Multi-agent performative prediction with greedy deployment and consensus seeking agents. Advances in Neural Information Processing Systems, 35:38449–38460, 2022.
  21. Stability and convergence of stochastic gradient clipping: Beyond lipschitz continuity and smoothness. In International Conference on Machine Learning, pages 7325–7335. PMLR, 2021.
  22. Learning differentially private recurrent language models. In International Conference on Learning Representations, 2018.
  23. Stochastic optimization for performative prediction. Advances in Neural Information Processing Systems, 33:4929–4939, 2020.
  24. Tomáš Mikolov et al. Statistical language models based on neural networks. Presentation at Google, Mountain View, 2nd April, 80(26), 2012.
  25. Outside the echo chamber: Optimizing the performative risk. In International Conference on Machine Learning, pages 7710–7720. PMLR, 2021.
  26. Ilya Mironov. Rényi differential privacy. In 2017 IEEE 30th computer security foundations symposium (CSF), pages 263–275. IEEE, 2017.
  27. Multiplayer performative prediction: Learning in decision-dependent games. Journal of Machine Learning Research, 24(202):1–56, 2023.
  28. On the difficulty of training recurrent neural networks. In International conference on machine learning, pages 1310–1318. Pmlr, 2013.
  29. Performative prediction. In International Conference on Machine Learning, pages 7599–7609. PMLR, 2020.
  30. Privacy-preserving deep learning. In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pages 1310–1321, 2015.
  31. Naum Zuselevich Shor. Minimization methods for non-differentiable functions, volume 3. Springer Science & Business Media, 2012.
  32. Differentially private distributed algorithms for stochastic aggregative games. Automatica, 142:110440, 2022.
  33. Differentially-private distributed algorithms for aggregative games with guaranteed convergence. IEEE Transactions on Automatic Control (Early access), pages 1–16, 2024. doi: 10.1109/TAC.2024.3351068.
  34. Federated learning with differential privacy: Algorithms and performance analysis. IEEE Transactions on Information Forensics and Security, 15:3454–3469, 2020.
  35. Differentially private distributed nash equilibrium seeking for aggregative games. IEEE Transactions on Automatic Control, 67(5):2451–2458, 2022.
  36. Why gradient clipping accelerates training: A theoretical justification for adaptivity. In International Conference on Learning Representations, 2020a.
  37. Why are adaptive methods good for attention models? Advances in Neural Information Processing Systems, 33:15383–15393, 2020b.
  38. Differentially private sgd without clipping bias: An error-feedback approach. In ICLR, 2024.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube