Task-Agnostic Privacy-Preserving Representation Learning for Federated Learning Against Attribute Inference Attacks (2312.06989v1)
Abstract: Federated learning (FL) has been widely studied recently due to its property to collaboratively train data from different devices without sharing the raw data. Nevertheless, recent studies show that an adversary can still be possible to infer private information about devices' data, e.g., sensitive attributes such as income, race, and sexual orientation. To mitigate the attribute inference attacks, various existing privacy-preserving FL methods can be adopted/adapted. However, all these existing methods have key limitations: they need to know the FL task in advance, or have intolerable computational overheads or utility losses, or do not have provable privacy guarantees. We address these issues and design a task-agnostic privacy-preserving presentation learning method for FL ({\bf TAPPFL}) against attribute inference attacks. TAPPFL is formulated via information theory. Specifically, TAPPFL has two mutual information goals, where one goal learns task-agnostic data representations that contain the least information about the private attribute in each device's data, and the other goal ensures the learnt data representations include as much information as possible about the device data to maintain FL utility. We also derive privacy guarantees of TAPPFL against worst-case attribute inference attacks, as well as the inherent tradeoff between utility preservation and privacy protection. Extensive results on multiple datasets and applications validate the effectiveness of TAPPFL to protect data privacy, maintain the FL utility, and be efficient as well. Experimental results also show that TAPPFL outperforms the existing defenses\footnote{Source code and full version: \url{https://github.com/TAPPFL}}.
- Deep Variational Information Bottleneck. In ICLR.
- Alibaba Federated Learning. 2022. https://federatedscope.io/.
- Privacy-preserving deep learning: Revisited and enhanced. In ATIS.
- Mutual information neural estimation. In ICML.
- Towards federated learning at scale: System design. MLSys.
- Practical secure aggregation for privacy-preserving machine learning. In CCS.
- Calabro, C. 2009. The exponential complexity of satisfiability problems. University of California, San Diego.
- Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In NIPS.
- CLUB: A Contrastive Log-ratio Upper Bound of Mutual Information. In ICML.
- Revealing and protecting labels in distributed training. In NeurIPS.
- Fully distributed privacy preserving mini-batch gradient descent learning. In IFIP DAIS.
- UCI Machine Learning Repository.
- Property inference attacks on fully connected neural networks using permutation invariant representations. In CCS.
- Differentially private federated learning: A client level perspective. arXiv.
- On choosing and bounding probability metrics. International statistical review, 70(3): 419–435.
- Generative adversarial nets. In NIPS.
- Google Federated Learning. 2022. https://federated.withgoogle.com/.
- Learning privately from multiparty data. In ICML.
- Equality of Opportunity in Supervised Learning.
- Learning deep representations by mutual information estimation and maximization. In ICLR.
- IBM Federated Learning. 2022. https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=models-federated-learning.
- AttriInfer: Inferring user attributes in online social networks using markov random fields. In WWW.
- Secure, privacy-preserving and federated machine learning in medical imaging. Nature Machine Intelligence.
- Lessons Learned from the Chameleon Testbed. In USENIX ATC.
- Training with the invisibles: Obfuscating images to share safely for learning visual recognition models. arXiv.
- An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4): 307–392.
- Krizhevsky, A. 2009. Learning multiple layers of features from tiny images. Technical report.
- Deepobfuscator: Adversarial training framework for privacy-preserving image classification. arXiv preprint arXiv:1909.04126.
- Federated learning: Challenges, methods, and future directions. IEEE Signal Processing Magazine.
- Information obfuscation of graph neural networks. In ICML.
- Privacy Adversarial Network: Representation Learning for Mobile Data Privacy. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 3(4): 1–18.
- Learning adversarially fair and transferable representations. In ICML.
- Communication-Efficient Learning of Deep Networks from Decentralized Data. In AISTATS.
- Learning Differentially Private Recurrent Language Models. In ICLR.
- Exploiting unintended feature leakage in collaborative learning. In IEEE SP.
- Microsoft Federated Learning. 2022. https://www.microsoft.com/en-us/research/blog/flute-a-scalable-federated-learning-simulation-platform/.
- R2DP: A Universal and Automated Approach to Optimizing the Randomization Mechanisms of Differential Privacy for Utility Metrics with No Known Optimal Distributions. In CCS, 677–696. ACM.
- Secureml: A system for scalable privacy-preserving machine learning. In IEEE SP.
- Invariant representations without adversarial training. In NeurIPS.
- Adversarial image perturbation for privacy protection a game theory perspective. In ICCV.
- Representation learning with contrastive predictive coding. arXiv.
- Deep private-feature extraction. IEEE TKDE.
- Multiparty differential privacy via aggregation of locally trained classifiers. In NIPS.
- Variational discriminator bottleneck: Improving imitation learning, inverse rl, and gans by constraining information flow. arXiv preprint arXiv:1810.00821.
- Learning privacy preserving encodings through adversarial training. In WACV.
- On variational bounds of mutual information. In ICML.
- The future of digital health with federated learning. NPJ digital medicine, 3(1): 119.
- Privacy-preserving deep learning. In CCS.
- Overlearning Reveals Sensitive Attributes. In ICLR.
- Learning controllable fair representations. In AISTATS.
- Visualizing data using t-SNE. JMLR.
- User-Level Label Leakage from Gradients in Federated Learning. In PTES.
- Privacy-Preserving Representation Learning on Graphs: A Mutual Information Perspective. In KDD.
- A Model-Agnostic Approach to Differentially Private Topic Mining. In KDD, 1835–1845. ACM.
- Federated learning with differential privacy: Algorithms and performance analysis. IEEE TIFS.
- Towards privacy-preserving visual recognition via adversarial training: A pilot study. In ECCV.
- Trade-offs and guarantees of adversarial representation learning for information obfuscation. In NeurIPS.
- Deep leakage from gradients. In NeurIPS.