QMGeo: Differentially Private Federated Learning via Stochastic Quantization with Mixed Truncated Geometric Distribution (2312.05761v2)
Abstract: Federated learning (FL) is a framework which allows multiple users to jointly train a global ML model by transmitting only model updates under the coordination of a parameter server, while being able to keep their datasets local. One key motivation of such distributed frameworks is to provide privacy guarantees to the users. However, preserving the users' datasets locally is shown to be not sufficient for privacy. Several differential privacy (DP) mechanisms have been proposed to provide provable privacy guarantees by introducing randomness into the framework, and majority of these mechanisms rely on injecting additive noise. FL frameworks also face the challenge of communication efficiency, especially as machine learning models grow in complexity and size. Quantization is a commonly utilized method, reducing the communication cost by transmitting compressed representation of the underlying information. Although there have been several studies on DP and quantization in FL, the potential contribution of the quantization method alone in providing privacy guarantees has not been extensively analyzed yet. We in this paper present a novel stochastic quantization method, utilizing a mixed geometric distribution to introduce the randomness needed to provide DP, without any additive noise. We provide convergence analysis for our framework and empirically study its performance.
- P. Kairouz and et al., “Advances and open problems in federated learning,” 2021.
- B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y. Arcas, “Communication-Efficient Learning of Deep Networks from Decentralized Data,” in Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, ser. Proceedings of Machine Learning Research, A. Singh and J. Zhu, Eds., vol. 54. PMLR, 20–22 Apr 2017, pp. 1273–1282. [Online]. Available: https://proceedings.mlr.press/v54/mcmahan17a.html
- M. M. Amiri, D. Gündüz, S. R. Kulkarni, and H. V. Poor, “Federated learning with quantized global model updates,” CoRR, vol. abs/2006.10672, 2020. [Online]. Available: https://arxiv.org/abs/2006.10672
- N. Shlezinger, M. Chen, Y. C. Eldar, H. V. Poor, and S. Cui, “UVeQFed: Universal vector quantization for federated learning,” IEEE Transactions on Signal Processing, vol. 69, pp. 500–514, 2021.
- M. Nasr, R. Shokri, and A. Houmansadr, “Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning,” in 2019 IEEE Symposium on Security and Privacy (SP). IEEE, may 2019. [Online]. Available: https://doi.org/10.1109%2Fsp.2019.00065
- R. Shokri, M. Stronati, C. Song, and V. Shmatikov, “Membership inference attacks against machine learning models,” 2017.
- S. Truex, L. Liu, M. E. Gursoy, L. Yu, and W. Wei, “Towards demystifying membership inference attacks,” ArXiv, vol. abs/1807.09173, 2018. [Online]. Available: https://api.semanticscholar.org/CorpusID:50778569
- C. Dwork, A. Roth et al., “The algorithmic foundations of differential privacy,” Foundations and Trends® in Theoretical Computer Science, vol. 9, no. 3–4, pp. 211–407, 2014.
- Ú. Erlingsson, A. Korolova, and V. Pihur, “RAPPOR: randomized aggregatable privacy-preserving ordinal response,” CoRR, vol. abs/1407.6981, 2014. [Online]. Available: http://arxiv.org/abs/1407.6981
- B. Ding, J. Kulkarni, and S. Yekhanin, “Collecting telemetry data privately,” CoRR, vol. abs/1712.01524, 2017. [Online]. Available: http://arxiv.org/abs/1712.01524
- J. Abowd, D. Kifer, S. L. Garfinkel, and A. Machanavajjhala, “Census topdown: Differentially private data, incremental schemas, and consistency with public knowledge,” 2019. [Online]. Available: https://api.semanticscholar.org/CorpusID:237407049
- N. Agarwal, A. T. Suresh, F. Yu, S. Kumar, and H. B. Mcmahan, “cpSGD: Communication-efficient and differentially-private distributed SGD,” 2018.
- P. Kairouz, Z. Liu, and T. Steinke, “The distributed discrete gaussian mechanism for federated learning with secure aggregation,” 2022.
- W.-N. Chen, A. Ozgur, and P. Kairouz, “The poisson binomial mechanism for unbiased federated learning with secure aggregation,” in International Conference on Machine Learning. PMLR, 2022, pp. 3490–3506.
- Y. Wang and T. Basar, “Quantization enabled privacy protection in decentralized stochastic optimization,” 2022.
- T. Olatayo, “Truncated geometric bootstrap method for time series stationary process,” Applied Mathematics, vol. 2014, 2014.
- B. Balle, G. Barthe, and M. Gaboardi, “Privacy amplification by subsampling: Tight analyses via couplings and divergences,” 2018.
- I. Mironov, “Rényi differential privacy,” in 2017 IEEE 30th computer security foundations symposium (CSF). IEEE, 2017, pp. 263–275.
- Y.-X. Wang, B. Balle, and S. P. Kasiviswanathan, “Subsampled rényi differential privacy and analytical moments accountant,” in The 22nd International Conference on Artificial Intelligence and Statistics. PMLR, 2019, pp. 1226–1235.
- S. Ghadimi and G. Lan, “Stochastic first-and zeroth-order methods for nonconvex stochastic programming,” SIAM Journal on Optimization, vol. 23, no. 4, pp. 2341–2368, 2013.
- Zixi Wang (15 papers)
- M. Cenk Gursoy (79 papers)