Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fine-Tuning Large Language Models with User-Level Differential Privacy (2407.07737v1)

Published 10 Jul 2024 in cs.LG, cs.CL, cs.CR, and cs.DC

Abstract: We investigate practical and scalable algorithms for training LLMs with user-level differential privacy (DP) in order to provably safeguard all the examples contributed by each user. We study two variants of DP-SGD with: (1) example-level sampling (ELS) and per-example gradient clipping, and (2) user-level sampling (ULS) and per-user gradient clipping. We derive a novel user-level DP accountant that allows us to compute provably tight privacy guarantees for ELS. Using this, we show that while ELS can outperform ULS in specific settings, ULS generally yields better results when each user has a diverse collection of examples. We validate our findings through experiments in synthetic mean estimation and LLM fine-tuning tasks under fixed compute budgets. We find that ULS is significantly better in settings where either (1) strong privacy guarantees are required, or (2) the compute budget is large. Notably, our focus on LLM-compatible training algorithms allows us to scale to models with hundreds of millions of parameters and datasets with hundreds of thousands of users.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zachary Charles (33 papers)
  2. Arun Ganesh (25 papers)
  3. Ryan McKenna (26 papers)
  4. H. Brendan McMahan (49 papers)
  5. Nicole Mitchell (7 papers)
  6. Krishna Pillutla (23 papers)
  7. Keith Rush (17 papers)
Citations (7)
X Twitter Logo Streamline Icon: https://streamlinehq.com