Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DP$^2$-VAE: Differentially Private Pre-trained Variational Autoencoders (2208.03409v2)

Published 5 Aug 2022 in cs.LG and cs.CR

Abstract: Modern machine learning systems achieve great success when trained on large datasets. However, these datasets usually contain sensitive information (e.g. medical records, face images), leading to serious privacy concerns. Differentially private generative models (DPGMs) emerge as a solution to circumvent such privacy concerns by generating privatized sensitive data. Similar to other differentially private (DP) learners, the major challenge for DPGM is also how to achieve a subtle balance between utility and privacy. We propose DP$2$-VAE, a novel training mechanism for variational autoencoders (VAE) with provable DP guarantees and improved utility via \emph{pre-training on private data}. Under the same DP constraints, DP$2$-VAE minimizes the perturbation noise during training, and hence improves utility. DP$2$-VAE is very flexible and easily amenable to many other VAE variants. Theoretically, we study the effect of pretraining on private data. Empirically, we conduct extensive experiments on image datasets to illustrate our superiority over baselines under various privacy budgets and evaluation metrics.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Dihong Jiang (2 papers)
  2. Guojun Zhang (43 papers)
  3. Mahdi Karami (11 papers)
  4. Xi Chen (1036 papers)
  5. Yunfeng Shao (34 papers)
  6. Yaoliang Yu (59 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.