Total Variation Meets Differential Privacy (2311.01553v2)
Abstract: The framework of approximate differential privacy is considered, and augmented by leveraging the notion of ``the total variation of a (privacy-preserving) mechanism'' (denoted by $\eta$-TV). With this refinement, an exact composition result is derived, and shown to be significantly tighter than the optimal bounds for differential privacy (which do not consider the total variation). Furthermore, it is shown that $(\varepsilon,\delta)$-DP with $\eta$-TV is closed under subsampling. The induced total variation of commonly used mechanisms are computed. Moreover, the notion of total variation of a mechanism is studied in the local privacy setting and privacy-utility tradeoffs are investigated. In particular, total variation distance and KL divergence are considered as utility functions and studied through the lens of contraction coefficients. Finally, the results are compared and connected to the locally differentially private setting.
- C. Dwork, F. McSherry, K. Nissim, and A. Smith, “Calibrating noise to sensitivity in private data analysis,” in Theory of Cryptography: 3rd Theory of Cryptography Conference, New York, NY, USA, Mar. 4-7, 2006. Springer, 2006, pp. 265–284.
- C. Dwork, K. Kenthapadi, F. McSherry, I. Mironov, and M. Naor, “Our data, ourselves: Privacy via distributed noise generation,” in Annual International Conference on the Theory and Applications of Cryptographic Techniques. Springer, 2006, pp. 486–503.
- I. Mironov, “Rényi differential privacy,” in IEEE 30th Computer Security Foundations Symposium. IEEE, 2017, pp. 263–275.
- C. Dwork and G. N. Rothblum, “Concentrated differential privacy,” arXiv preprint arXiv:1603.01887, 2016.
- J. Dong, A. Roth, and W. J. Su, “Gaussian differential privacy,” 2019. [Online]. Available: https://arxiv.org/abs/1905.02383
- L. Wasserman and S. Zhou, “A statistical framework for differential privacy,” Journal of the American Statistical Association, vol. 105, no. 489, pp. 375–389, 2010.
- P. Kairouz, S. Oh, and P. Viswanath, “The composition theorem for differential privacy,” in International conference on machine learning. PMLR, 2015, pp. 1376–1385.
- Q. Geng, P. Kairouz, S. Oh, and P. Viswanath, “The staircase mechanism in differential privacy,” IEEE Journal of Selected Topics in Signal Processing, vol. 9, no. 7, pp. 1176–1184, 2015.
- J. C. Duchi, M. I. Jordan, and M. J. Wainwright, “Local privacy, data processing inequalities, and statistical minimax rates,” 2014.
- P. Kairouz, S. Oh, and P. Viswanath, “Extremal mechanisms for local differential privacy,” vol. 17, no. 17, 2016, pp. 1–51.
- S. Asoodeh and H. Zhang, “Contraction of locally differentially private mechanisms,” arXiv preprint arXiv:2210.13386, 2022.
- K. Chatzikokolakis, G. Cherubin, C. Palamidessi, and C. Troncoso, “Bayes security: A not so average metric,” in IEEE 36th Computer Security Foundations Symposium, 2023, pp. 388–406.
- R. F. Barber and J. C. Duchi, “Privacy and statistical risk: Formalisms and minimax bounds,” 2014.
- Q. Geng, W. Ding, R. Guo, and S. Kumar, “Optimal noise-adding mechanism in additive differential privacy,” in The 22nd International Conference on Artificial Intelligence and Statistics. PMLR, 2019, pp. 11–20.
- J. Jia, C. Tan, Z. Liu, X. Li, Z. Liu, S. Lv, and C. Dong, “Total variation distance privacy: Accurately measuring inference attacks and improving utility,” Information Sciences, vol. 626, pp. 537–558, 2023.
- C. Dwork, M. Hardt, T. Pitassi, O. Reingold, and R. Zemel, “Fairness through awareness,” in Proceedings of the 3rd Innovations in Theoretical Computer Science Conference. New York, NY, USA: Association for Computing Machinery, 2012, p. 214–226.
- R. Bassily, K. Nissim, A. Smith, T. Steinke, U. Stemmer, and J. Ullman, “Algorithmic stability for adaptive data analysis,” in Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing. New York, NY, USA: Association for Computing Machinery, 2016, p. 1046–1059.
- M. Raginsky, A. Rakhlin, M. Tsao, Y. Wu, and A. Xu, “Information-theoretic analysis of stability and bias of learning algorithms,” in IEEE Information Theory Workshop, 2016.
- M. Yaghini, B. Kulynych, and C. Troncoso, “Disparate vulnerability: on the unfairness of privacy attacks against machine learning,” CoRR, vol. abs/1906.00389, 2019.
- B. Kulynych, Y.-Y. Yang, Y. Yu, J. Błasiok, and P. Nakkiran, “What you see is what you get: Principled deep learning via distributional generalization,” vol. 35, 2022, pp. 2168–2183.
- D. Blackwell, “Equivalent comparisons of experiments,” The annals of mathematical statistics, pp. 265–272, 1953.
- T. M. Cover and J. A. Thomas, “Elements of information theory 2nd edition,” Willey-Interscience: NJ, 2006.
- S. P. Kasiviswanathan, H. K. Lee, K. Nissim, S. Raskhodnikova, and A. Smith, “What can we learn privately?” SIAM Journal on Computing, vol. 40, no. 3, pp. 793–826, 2011.
- A. Smith, “Differential privacy and the secrecy of the sample,” Sep 2009. [Online]. Available: https://adamdsmith.wordpress.com/2009/09/02/sample-secrecy/
- K. Chaudhuri and N. Mishra, “When random sampling preserves privacy,” in Annual International Cryptology Conference. Springer, 2006, pp. 198–213.
- B. Balle, G. Barthe, and M. Gaboardi, “Privacy amplification by subsampling: Tight analyses via couplings and divergences,” Advances in neural information processing systems, vol. 31, 2018.
- T. Steinke, “Composition of differential privacy & privacy amplification by subsampling,” arXiv preprint arXiv:2210.00597, 2022.
- A. Koskela, J. Jälkö, L. Prediger, and A. Honkela, “Tight differential privacy for discrete-valued mechanisms and for the subsampled gaussian mechanism using FFT,” in International Conference on Artificial Intelligence and Statistics. PMLR, 2021, pp. 3358–3366.
- S. Gopi, Y. T. Lee, and L. Wutschitz, “Numerical composition of differential privacy,” vol. 34, 2021, pp. 11 631–11 642.
- Y. Zhu, J. Dong, and Y.-X. Wang, “Optimal accounting of differential privacy via characteristic function,” in International Conference on Artificial Intelligence and Statistics. PMLR, 2022, pp. 4782–4817.
- B. Balle and Y.-X. Wang, “Improving the gaussian mechanism for differential privacy: Analytical calibration and optimal denoising,” in International Conference on Machine Learning. PMLR, 2018, pp. 394–403.
- Q. Geng and P. Viswanath, “The optimal mechanism in differential privacy,” in 2014 IEEE International Symposium on Information Theory, 2014, pp. 2371–2375.
- M. Abadi, A. Chu, I. Goodfellow, H. B. McMahan, I. Mironov, K. Talwar, and L. Zhang, “Deep learning with differential privacy,” in Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, 2016, pp. 308–318.
- R. L. Dobrushin, “Central limit theorem for nonstationary markov chains. i,” Theory of Probability & Its Applications, vol. 1, no. 1, pp. 65–80, 1956.
- S. Asoodeh, M. Aliakbarpour, and F. P. Calmon, “Local differential privacy is equivalent to contraction of an f𝑓fitalic_f-divergence,” in IEEE International Symposium on Information Theory. IEEE, 2021, pp. 545–550.
- B. Zamanlooy and S. Asoodeh, “Strong data processing inequalities for locally differentially private mechanisms,” in IEEE International Symposium on Information Theory. IEEE, 2023, pp. 1794–1799.
- M.-D. Choi, M. B. Ruskai, and E. Seneta, “Equivalence of certain entropy contraction coefficients,” Linear algebra and its applications, vol. 208, pp. 29–36, 1994.
- O. Ordentlich and Y. Polyanskiy, “Strong data processing constant is achieved by binary inputs,” IEEE Transactions on Information Theory, vol. 68, no. 3, pp. 1480–1481, 2021.
- Y. Polyanskiy and Y. Wu, “Strong data-processing inequalities for channels and bayesian networks,” in Convexity and Concentration. Springer, 2017, pp. 211–249.
- L. Györfi and I. Vajda, “A class of modified pearson and neyman statistics,” Statistics & Risk Modeling, vol. 19, no. 3, pp. 239–252, 2001.
- M. Raginsky, “Strong data processing inequalities and ΦΦ\Phiroman_Φ-sobolev inequalities for discrete channels,” IEEE Transactions on Information Theory, vol. 62, no. 6, pp. 3355–3389, 2016.
- I. Sason and S. Verdú, “f𝑓fitalic_f-divergence inequalities,” IEEE Transactions on Information Theory, vol. 62, no. 11, pp. 5973–6006, 2016.
- J. C. Duchi and F. Ruan, “The right complexity measure in locally private estimation: It is not the fisher information,” arXiv preprint arXiv:1806.05756, 2018.