Exploring Privacy and Fairness Risks in Sharing Diffusion Models: An Adversarial Perspective (2402.18607v3)
Abstract: Diffusion models have recently gained significant attention in both academia and industry due to their impressive generative performance in terms of both sampling quality and distribution coverage. Accordingly, proposals are made for sharing pre-trained diffusion models across different organizations, as a way of improving data utilization while enhancing privacy protection by avoiding sharing private data directly. However, the potential risks associated with such an approach have not been comprehensively examined. In this paper, we take an adversarial perspective to investigate the potential privacy and fairness risks associated with the sharing of diffusion models. Specifically, we investigate the circumstances in which one party (the sharer) trains a diffusion model using private data and provides another party (the receiver) black-box access to the pre-trained model for downstream tasks. We demonstrate that the sharer can execute fairness poisoning attacks to undermine the receiver's downstream models by manipulating the training data distribution of the diffusion model. Meanwhile, the receiver can perform property inference attacks to reveal the distribution of sensitive features in the sharer's dataset. Our experiments conducted on real-world datasets demonstrate remarkable attack performance on different types of diffusion models, which highlights the critical importance of robust data auditing and privacy protection protocols in pertinent applications.
- R. H. Hariri, E. M. Fredericks, and K. M. Bowers, “Uncertainty in big data analytics: survey, opportunities, and challenges,” J. Big Data, vol. 6, p. 44, 2019.
- R. Hamilton and W. A. Sodeman, “The questions we ask: Opportunities and challenges for using big data analytics to strategically manage human capital resources,” Business Horizons, vol. 63, no. 1, pp. 85–95, 2020.
- S. M. Ayyoubzadeh, S. M. Ayyoubzadeh, and M. Esmaeili, “Clinical data sharing using generative adversarial networks,” Connected Health, vol. 1, no. 3, pp. 98–100, 2022.
- P. A. Moghadam, S. V. Dalen, K. C. Martin, J. K. Lennerz, S. Yip, H. Farahani, and A. Bashashati, “A morphology focused diffusion probabilistic model for synthesis of histopathology images,” in IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2023, Waikoloa, HI, USA, January 2-7, 2023. IEEE, 2023, pp. 1999–2008.
- Z. Szafranowska, R. Osuala, B. Breier, K. Kushibar, K. Lekadir, and O. Díaz, “Sharing generative models instead of private data: A simulation study on mammography patch classification,” CoRR, vol. abs/2203.04961, 2022.
- Q. Chang, H. Qu, Y. Zhang, M. Sabuncu, C. Chen, T. Zhang, and D. N. Metaxas, “Synthetic learning: Learn from distributed asynchronized discriminator gan without sharing medical image data,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 13 856–13 866.
- T. Han, S. Nebelung, C. Haarburger, N. Horst, S. Reinartz, D. Merhof, F. Kiessling, V. Schulz, and D. Truhn, “Breaking medical data sharing boundaries by using synthesized radiographs,” Science advances, vol. 6, no. 49, p. eabb7973, 2020.
- H. He, S. Zhao, Y. Xi, and J. C. Ho, “Meddiff: Generating electronic health records using accelerated denoising diffusion model,” CoRR, vol. abs/2302.04355, 2023.
- J. Shipard, A. Wiliem, K. N. Thanh, W. Xiang, and C. Fookes, “Diversity is definitely needed: Improving model-agnostic zero-shot classification via stable diffusion,” 2023.
- R. He, S. Sun, X. Yu, C. Xue, W. Zhang, P. H. S. Torr, S. Bai, and X. Qi, “Is synthetic data from generative models ready for image recognition?” CoRR, vol. abs/2210.07574, 2022. [Online]. Available: https://doi.org/10.48550/arXiv.2210.07574
- A. Roy, A. Shah, K. Shah, A. Roy, and R. Chellappa, “Diffalign : Few-shot learning using diffusion based synthesis and alignment,” CoRR, vol. abs/2212.05404, 2022.
- A. Kotelnikov, D. Baranchuk, I. Rubachev, and A. Babenko, “Tabddpm: Modelling tabular data with diffusion models,” CoRR, vol. abs/2209.15421, 2022.
- J. Ho, A. Jain, and P. Abbeel, “Denoising diffusion probabilistic models,” in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020.
- Y. Song and S. Ermon, “Generative modeling by estimating gradients of the data distribution,” Advances in neural information processing systems, vol. 32, 2019.
- Y. Song, J. Sohl-Dickstein, D. P. Kingma, A. Kumar, S. Ermon, and B. Poole, “Score-based generative modeling through stochastic differential equations,” in 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021.
- H. Bansal and A. Grover, “Leaving reality to imagination: Robust classification via generated datasets,” CoRR, vol. abs/2302.02503, 2023.
- S. Bond-Taylor, A. Leach, Y. Long, and C. G. Willcocks, “Deep generative modelling: A comparative review of vaes, gans, normalizing flows, energy-based and autoregressive models,” IEEE transactions on pattern analysis and machine intelligence, 2021.
- Y. Qu, S. Yu, W. Zhou, and Y. Tian, “Gan-driven personalized spatial-temporal private data sharing in cyber-physical social systems,” IEEE Trans. Netw. Sci. Eng., vol. 7, no. 4, pp. 2576–2586, 2020.
- R. Rombach, A. Blattmann, D. Lorenz, P. Esser, and B. Ommer, “High-resolution image synthesis with latent diffusion models,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022. IEEE, 2022, pp. 10 674–10 685.
- F. Tramèr, F. Zhang, A. Juels, M. K. Reiter, and T. Ristenpart, “Stealing machine learning models via prediction apis,” in 25th USENIX Security Symposium, USENIX Security 16, Austin, TX, USA, August 10-12, 2016. USENIX Association, 2016, pp. 601–618. [Online]. Available: https://www.usenix.org/conference/usenixsecurity16/technical-sessions/presentation/tramer
- N. Carlini, J. Hayes, M. Nasr, M. Jagielski, V. Sehwag, F. Tramer, B. Balle, D. Ippolito, and E. Wallace, “Extracting training data from diffusion models,” in 32nd USENIX Security Symposium (USENIX Security 23), 2023, pp. 5253–5270.
- J. Duan, F. Kong, S. Wang, X. Shi, and K. Xu, “Are diffusion models vulnerable to membership inference attacks?” CoRR, vol. abs/2302.01316, 2023.
- H. Hu and J. Pang, “Membership inference of diffusion models,” CoRR, vol. abs/2301.09956, 2023. [Online]. Available: https://doi.org/10.48550/arXiv.2301.09956
- M. Nasr, R. Shokri, and A. Houmansadr, “Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning,” in 2019 IEEE Symposium on Security and Privacy, SP 2019, San Francisco, CA, USA, May 19-23, 2019. IEEE, 2019, pp. 739–753.
- A. Salem, Y. Zhang, M. Humbert, P. Berrang, M. Fritz, and M. Backes, “Ml-leaks: Model and data independent membership inference attacks and defenses on machine learning models,” in 26th Annual Network and Distributed System Security Symposium, NDSS 2019, San Diego, California, USA, February 24-27, 2019. The Internet Society, 2019.
- J. Zhou, Y. Chen, C. Shen, and Y. Zhang, “Property inference attacks against gans,” in 29th Annual Network and Distributed System Security Symposium, NDSS 2022, San Diego, California, USA, April 24-28, 2022. The Internet Society, 2022.
- S. Mahloujifar, E. Ghosh, and M. Chase, “Property inference from poisoning,” in 43rd IEEE Symposium on Security and Privacy, SP 2022, San Francisco, CA, USA, May 22-26, 2022. IEEE, 2022, pp. 1120–1137.
- N. Konstantinov and C. H. Lampert, “On the impossibility of fairness-aware learning from corrupted data,” in Algorithmic Fairness through the Lens of Causality and Robustness workshop. PMLR, 2022, pp. 59–83.
- H. Chang, T. D. Nguyen, S. K. Murakonda, E. Kazemi, and R. Shokri, “On adversarial bias and the robustness of fair machine learning,” CoRR, vol. abs/2006.08669, 2020. [Online]. Available: https://arxiv.org/abs/2006.08669
- M. Van, W. Du, X. Wu, and A. Lu, “Poisoning attacks on fair machine learning,” in Database Systems for Advanced Applications - 27th International Conference, DASFAA 2022, Virtual Event, April 11-14, 2022, Proceedings, Part I, ser. Lecture Notes in Computer Science, vol. 13245. Springer, 2022, pp. 370–386.
- D. Solans, B. Biggio, and C. Castillo, “Poisoning attacks on algorithmic fairness,” in Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 2020, pp. 162–177.
- C. Jo, J. Sohn, and K. Lee, “Breaking fair binary classification with optimal flipping attacks,” in IEEE International Symposium on Information Theory, ISIT 2022, Espoo, Finland, June 26 - July 1, 2022. IEEE, 2022, pp. 1453–1458.
- Z. Wang, Y. Huang, M. Song, L. Wu, F. Xue, and K. Ren, “Poisoning-assisted property inference attack against federated learning,” IEEE Transactions on Dependable and Secure Computing, 2022.
- L. Melis, C. Song, E. De Cristofaro, and V. Shmatikov, “Exploiting unintended feature leakage in collaborative learning,” in 2019 IEEE symposium on security and privacy (SP). IEEE, 2019, pp. 691–706.
- K. Ganju, Q. Wang, W. Yang, C. A. Gunter, and N. Borisov, “Property inference attacks on fully connected neural networks using permutation invariant representations,” in Proceedings of the 2018 ACM SIGSAC conference on computer and communications security, 2018, pp. 619–633.
- Y. Song and S. Ermon, “Improved techniques for training score-based generative models,” Advances in neural information processing systems, vol. 33, pp. 12 438–12 448, 2020.
- A. Agarwal, A. Beygelzimer, M. Dudík, J. Langford, and H. Wallach, “A reductions approach to fair classification,” in International conference on machine learning. PMLR, 2018, pp. 60–69.
- M. Hardt, E. Price, and N. Srebro, “Equality of opportunity in supervised learning,” in Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, 2016, pp. 3315–3323.
- A. Ghassami, S. Khodadadian, and N. Kiyavash, “Fairness in supervised learning: An information theoretic approach,” in 2018 IEEE International Symposium on Information Theory, ISIT 2018, Vail, CO, USA, June 17-22, 2018. IEEE, 2018, pp. 176–180.
- N. Carlini, S. Chien, M. Nasr, S. Song, A. Terzis, and F. Tramèr, “Membership inference attacks from first principles,” in 43rd IEEE Symposium on Security and Privacy, SP 2022, San Francisco, CA, USA, May 22-26, 2022. IEEE, 2022, pp. 1897–1914.
- D. Acemoglu, A. Makhdoumi, A. Malekian, and A. Ozdaglar, “Too much data: Prices and inefficiencies in data markets,” American Economic Journal: Microeconomics, vol. 14, no. 4, pp. 218–256, 2022.
- H. Chang and R. Shokri, “Bias propagation in federated learning,” in The Eleventh International Conference on Learning Representations, 2023.
- X. Luo, Y. Wu, X. Xiao, and B. C. Ooi, “Feature inference attack on model predictions in vertical federated learning,” in 37th IEEE International Conference on Data Engineering, ICDE 2021, Chania, Greece, April 19-22, 2021. IEEE, 2021, pp. 181–192.
- J. Kang, T. Xie, X. Wu, R. Maciejewski, and H. Tong, “Infofair: Information-theoretic intersectional fairness,” in IEEE International Conference on Big Data, Big Data 2022, Osaka, Japan, December 17-20, 2022. IEEE, 2022, pp. 1455–1464.
- A. M. Fraser and H. L. Swinney, “Independent coordinates for strange attractors from mutual information,” Physical review A, vol. 33, no. 2, p. 1134, 1986.
- A. Kraskov, H. Stögbauer, and P. Grassberger, “Estimating mutual information,” Physical review E, vol. 69, no. 6, p. 066138, 2004.
- M. I. Belghazi, A. Baratin, S. Rajeshwar, S. Ozair, Y. Bengio, A. Courville, and D. Hjelm, “Mutual information neural estimation,” in International conference on machine learning. PMLR, 2018, pp. 531–540.
- Y. Chen, M. Mancini, X. Zhu, and Z. Akata, “Semi-supervised and unsupervised deep visual learning: A survey,” CoRR, vol. abs/2208.11296, 2022. [Online]. Available: https://doi.org/10.48550/arXiv.2208.11296
- A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark et al., “Learning transferable visual models from natural language supervision,” in International conference on machine learning. PMLR, 2021, pp. 8748–8763.
- Z. Niu, M. Zhou, L. Wang, X. Gao, and G. Hua, “Ordinal regression with multiple output CNN for age estimation,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016. IEEE Computer Society, 2016, pp. 4920–4928.
- W. Hoeffding, “Probability inequalities for sums of bounded random variables,” Journal of the American Statistical Association, vol. 58, no. 301, pp. 13–30, 1963.
- R. Ragonesi, P. Morerio, and V. Murino, “Learning unbiased classifiers from biased data with meta-learning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 1–9.
- C. Reimers, P. Bodesheim, J. Runge, and J. Denzler, “Towards learning an unbiased classifier from biased data via conditional adversarial debiasing,” arXiv preprint arXiv:2103.06179, 2021.
- H. Jiang and O. Nachum, “Identifying and correcting label bias in machine learning,” in The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy], ser. Proceedings of Machine Learning Research, vol. 108. PMLR, 2020, pp. 702–712.
- P. Dhariwal and A. Nichol, “Diffusion models beat gans on image synthesis,” Advances in neural information processing systems, vol. 34, pp. 8780–8794, 2021.
- M. Galar, A. Fernández, E. Barrenechea, H. Bustince, and F. Herrera, “An overview of ensemble methods for binary classifiers in multi-class problems: Experimental study on one-vs-one and one-vs-all schemes,” Pattern Recognition, vol. 44, no. 8, pp. 1761–1776, 2011.
- T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, S. Agarwal, A. Herbert-Voss, G. Krueger, T. Henighan, R. Child, A. Ramesh, D. M. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin, S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever, and D. Amodei, “Language models are few-shot learners,” CoRR, vol. abs/2005.14165, 2020.
- K. Sohn, D. Berthelot, N. Carlini, Z. Zhang, H. Zhang, C. A. Raffel, E. D. Cubuk, A. Kurakin, and C.-L. Li, “Fixmatch: Simplifying semi-supervised learning with consistency and confidence,” Advances in neural information processing systems, vol. 33, pp. 596–608, 2020.
- Y. LeCun and C. Cortes, “MNIST handwritten digit database,” http://yann.lecun.com/exdb/mnist/, 2010.
- Z. Liu, P. Luo, X. Wang, and X. Tang, “Deep learning face attributes in the wild,” in Proceedings of International Conference on Computer Vision (ICCV), December 2015.
- D. Dua and C. Graff, “UCI machine learning repository,” 2017. [Online]. Available: http://archive.ics.uci.edu/ml
- T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida, “Spectral normalization for generative adversarial networks,” arXiv preprint arXiv:1802.05957, 2018.
- T. Chen, X. Zhai, M. Ritter, M. Lucic, and N. Houlsby, “Self-supervised gans via auxiliary rotation loss,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 12 154–12 163.
- C. Chadebec, E. Thibeau-Sutre, N. Burgos, and S. Allassonnière, “Data augmentation in high dimensional low sample size setting using a geometry-based variational autoencoder,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 3, pp. 2879–2896, 2022.
- X. Luo, Y. Jiang, and X. Xiao, “Feature inference attack on shapley values,” in Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, CCS 2022, Los Angeles, CA, USA, November 7-11, 2022. ACM, 2022, pp. 2233–2247.
- C. Shorten and T. M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” J. Big Data, vol. 6, p. 60, 2019.
- Wikipedia, “Principal component analysis,” https://en.wikipedia.org/wiki/Principal_component_analysis, 2023, accessed: 2023-04-30.
- X. Cao, M. Fang, J. Liu, and N. Z. Gong, “Fltrust: Byzantine-robust federated learning via trust bootstrapping,” in 28th Annual Network and Distributed System Security Symposium, NDSS 2021, virtually, February 21-25, 2021. The Internet Society, 2021.
- M. Fang, X. Cao, J. Jia, and N. Gong, “Local model poisoning attacks to {{\{{Byzantine-Robust}}\}} federated learning,” in 29th USENIX security symposium (USENIX Security 20), 2020, pp. 1605–1622.
- L. Dinh, D. Krueger, and Y. Bengio, “NICE: non-linear independent components estimation,” in 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Workshop Track Proceedings, 2015.
- A. van den Oord, N. Kalchbrenner, and K. Kavukcuoglu, “Pixel recurrent neural networks,” in Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016, ser. JMLR Workshop and Conference Proceedings, vol. 48. JMLR.org, 2016, pp. 1747–1756.
- D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” in 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014.
- I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” Communications of the ACM, vol. 63, no. 11, pp. 139–144, 2020.
- R. Shokri, M. Stronati, C. Song, and V. Shmatikov, “Membership inference attacks against machine learning models,” in IEEE S&\&&P, 2017, pp. 3–18.
- X. Luo, X. Xiao, Y. Wu, J. Liu, and B. C. Ooi, “A fusion-denoising attack on instahide with data augmentation,” in Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Virtual Event, February 22 - March 1, 2022. AAAI Press, 2022, pp. 1899–1907.
- B. Hitaj, G. Ateniese, and F. Perez-Cruz, “Deep models under the gan: information leakage from collaborative deep learning,” in Proceedings of the 2017 ACM SIGSAC conference on computer and communications security, 2017, pp. 603–618.
- M. Fredrikson, S. Jha, and T. Ristenpart, “Model inversion attacks that exploit confidence information and basic countermeasures,” in Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, October 12-16, 2015. ACM, 2015, pp. 1322–1333.
- X. Chen, C. Liu, B. Li, K. Lu, and D. Song, “Targeted backdoor attacks on deep learning systems using data poisoning,” CoRR, vol. abs/1712.05526, 2017.
- E. Bagdasaryan, A. Veit, Y. Hua, D. Estrin, and V. Shmatikov, “How to backdoor federated learning,” in International conference on artificial intelligence and statistics. PMLR, 2020, pp. 2938–2948.
- M. Alberti, V. Pondenkandath, M. Wursch, M. Bouillon, M. Seuret, R. Ingold, and M. Liwicki, “Are you tampering with my data?” in Proceedings of the European Conference on Computer Vision (ECCV) Workshops, 2018, pp. 0–0.
- W. Chen, D. Song, and B. Li, “Trojdiff: Trojan attacks on diffusion models with diverse targets,” CoRR, vol. abs/2303.05762, 2023.
- L. Muñoz-González, B. Pfitzner, M. Russo, J. Carnerero-Cano, and E. C. Lupu, “Poisoning attacks with generative adversarial nets,” CoRR, vol. abs/1906.07773, 2019.
- X. Zhang, X. Zhu, and L. Lessard, “Online data poisoning attacks,” in Learning for Dynamics and Control. PMLR, 2020, pp. 201–210.
- Y. Yu, X. Gao, and C. Xu, “LAFEAT: piercing through adversarial defenses with latent features,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, June 19-25, 2021. Computer Vision Foundation / IEEE, 2021, pp. 5735–5745.
- L. Muñoz-González, B. Biggio, A. Demontis, A. Paudice, V. Wongrassamee, E. C. Lupu, and F. Roli, “Towards poisoning of deep learning algorithms with back-gradient optimization,” in Proceedings of the 10th ACM workshop on artificial intelligence and security, 2017, pp. 27–38.
- N. Mehrabi, M. Naveed, F. Morstatter, and A. Galstyan, “Exacerbating algorithmic bias through fairness attacks,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 10, 2021, pp. 8930–8938.
- J. Jordon, J. Yoon, and M. van der Schaar, “PATE-GAN: generating synthetic data with differential privacy guarantees,” in 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
- S. Ghalebikesabi, L. Berrada, S. Gowal, I. Ktena, R. Stanforth, J. Hayes, S. De, S. L. Smith, O. Wiles, and B. Balle, “Differentially private diffusion models generate useful synthetic images,” CoRR, vol. abs/2302.13861, 2023.
- X. Xiao, G. Wang, and J. Gehrke, “Differential privacy via wavelet transforms,” IEEE Trans. Knowl. Data Eng., vol. 23, no. 8, pp. 1200–1214, 2011.
- R. C. Geyer, T. Klein, and M. Nabi, “Differentially private federated learning: A client level perspective,” CoRR, vol. abs/1712.07557, 2017.
- A. Torralba and A. A. Efros, “Unbiased look at dataset bias,” in The 24th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011, Colorado Springs, CO, USA, 20-25 June 2011. IEEE Computer Society, 2011, pp. 1521–1528.