Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DP-TBART: A Transformer-based Autoregressive Model for Differentially Private Tabular Data Generation (2307.10430v1)

Published 19 Jul 2023 in cs.LG and cs.CR

Abstract: The generation of synthetic tabular data that preserves differential privacy is a problem of growing importance. While traditional marginal-based methods have achieved impressive results, recent work has shown that deep learning-based approaches tend to lag behind. In this work, we present Differentially-Private TaBular AutoRegressive Transformer (DP-TBART), a transformer-based autoregressive model that maintains differential privacy and achieves performance competitive with marginal-based methods on a wide variety of datasets, capable of even outperforming state-of-the-art methods in certain settings. We also provide a theoretical framework for understanding the limitations of marginal-based approaches and where deep learning-based approaches stand to contribute most. These results suggest that deep learning-based techniques should be considered as a viable alternative to marginal-based methods in the generation of differentially private synthetic tabular data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Rodrigo Castellon (4 papers)
  2. Achintya Gopal (13 papers)
  3. Brian Bloniarz (1 paper)
  4. David Rosenberg (12 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.