Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

QMGeo: Differentially Private Federated Learning via Stochastic Quantization with Mixed Truncated Geometric Distribution (2312.05761v2)

Published 10 Dec 2023 in cs.LG and cs.AI

Abstract: Federated learning (FL) is a framework which allows multiple users to jointly train a global ML model by transmitting only model updates under the coordination of a parameter server, while being able to keep their datasets local. One key motivation of such distributed frameworks is to provide privacy guarantees to the users. However, preserving the users' datasets locally is shown to be not sufficient for privacy. Several differential privacy (DP) mechanisms have been proposed to provide provable privacy guarantees by introducing randomness into the framework, and majority of these mechanisms rely on injecting additive noise. FL frameworks also face the challenge of communication efficiency, especially as machine learning models grow in complexity and size. Quantization is a commonly utilized method, reducing the communication cost by transmitting compressed representation of the underlying information. Although there have been several studies on DP and quantization in FL, the potential contribution of the quantization method alone in providing privacy guarantees has not been extensively analyzed yet. We in this paper present a novel stochastic quantization method, utilizing a mixed geometric distribution to introduce the randomness needed to provide DP, without any additive noise. We provide convergence analysis for our framework and empirically study its performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. P. Kairouz and et al., “Advances and open problems in federated learning,” 2021.
  2. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y. Arcas, “Communication-Efficient Learning of Deep Networks from Decentralized Data,” in Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, ser. Proceedings of Machine Learning Research, A. Singh and J. Zhu, Eds., vol. 54.   PMLR, 20–22 Apr 2017, pp. 1273–1282. [Online]. Available: https://proceedings.mlr.press/v54/mcmahan17a.html
  3. M. M. Amiri, D. Gündüz, S. R. Kulkarni, and H. V. Poor, “Federated learning with quantized global model updates,” CoRR, vol. abs/2006.10672, 2020. [Online]. Available: https://arxiv.org/abs/2006.10672
  4. N. Shlezinger, M. Chen, Y. C. Eldar, H. V. Poor, and S. Cui, “UVeQFed: Universal vector quantization for federated learning,” IEEE Transactions on Signal Processing, vol. 69, pp. 500–514, 2021.
  5. M. Nasr, R. Shokri, and A. Houmansadr, “Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning,” in 2019 IEEE Symposium on Security and Privacy (SP).   IEEE, may 2019. [Online]. Available: https://doi.org/10.1109%2Fsp.2019.00065
  6. R. Shokri, M. Stronati, C. Song, and V. Shmatikov, “Membership inference attacks against machine learning models,” 2017.
  7. S. Truex, L. Liu, M. E. Gursoy, L. Yu, and W. Wei, “Towards demystifying membership inference attacks,” ArXiv, vol. abs/1807.09173, 2018. [Online]. Available: https://api.semanticscholar.org/CorpusID:50778569
  8. C. Dwork, A. Roth et al., “The algorithmic foundations of differential privacy,” Foundations and Trends® in Theoretical Computer Science, vol. 9, no. 3–4, pp. 211–407, 2014.
  9. Ú. Erlingsson, A. Korolova, and V. Pihur, “RAPPOR: randomized aggregatable privacy-preserving ordinal response,” CoRR, vol. abs/1407.6981, 2014. [Online]. Available: http://arxiv.org/abs/1407.6981
  10. B. Ding, J. Kulkarni, and S. Yekhanin, “Collecting telemetry data privately,” CoRR, vol. abs/1712.01524, 2017. [Online]. Available: http://arxiv.org/abs/1712.01524
  11. J. Abowd, D. Kifer, S. L. Garfinkel, and A. Machanavajjhala, “Census topdown: Differentially private data, incremental schemas, and consistency with public knowledge,” 2019. [Online]. Available: https://api.semanticscholar.org/CorpusID:237407049
  12. N. Agarwal, A. T. Suresh, F. Yu, S. Kumar, and H. B. Mcmahan, “cpSGD: Communication-efficient and differentially-private distributed SGD,” 2018.
  13. P. Kairouz, Z. Liu, and T. Steinke, “The distributed discrete gaussian mechanism for federated learning with secure aggregation,” 2022.
  14. W.-N. Chen, A. Ozgur, and P. Kairouz, “The poisson binomial mechanism for unbiased federated learning with secure aggregation,” in International Conference on Machine Learning.   PMLR, 2022, pp. 3490–3506.
  15. Y. Wang and T. Basar, “Quantization enabled privacy protection in decentralized stochastic optimization,” 2022.
  16. T. Olatayo, “Truncated geometric bootstrap method for time series stationary process,” Applied Mathematics, vol. 2014, 2014.
  17. B. Balle, G. Barthe, and M. Gaboardi, “Privacy amplification by subsampling: Tight analyses via couplings and divergences,” 2018.
  18. I. Mironov, “Rényi differential privacy,” in 2017 IEEE 30th computer security foundations symposium (CSF).   IEEE, 2017, pp. 263–275.
  19. Y.-X. Wang, B. Balle, and S. P. Kasiviswanathan, “Subsampled rényi differential privacy and analytical moments accountant,” in The 22nd International Conference on Artificial Intelligence and Statistics.   PMLR, 2019, pp. 1226–1235.
  20. S. Ghadimi and G. Lan, “Stochastic first-and zeroth-order methods for nonconvex stochastic programming,” SIAM Journal on Optimization, vol. 23, no. 4, pp. 2341–2368, 2013.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Zixi Wang (15 papers)
  2. M. Cenk Gursoy (79 papers)
Citations (1)