Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Benchmarking Differential Privacy and Federated Learning for BERT Models (2106.13973v2)

Published 26 Jun 2021 in cs.CL, cs.CR, and cs.LG

Abstract: NLP techniques can be applied to help with the diagnosis of medical conditions such as depression, using a collection of a person's utterances. Depression is a serious medical illness that can have adverse effects on how one feels, thinks, and acts, which can lead to emotional and physical problems. Due to the sensitive nature of such data, privacy measures need to be taken for handling and training models with such data. In this work, we study the effects that the application of Differential Privacy (DP) has, in both a centralized and a Federated Learning (FL) setup, on training contextualized LLMs (BERT, ALBERT, RoBERTa and DistilBERT). We offer insights on how to privately train NLP models and what architectures and setups provide more desirable privacy utility trade-offs. We envisage this work to be used in future healthcare and mental health studies to keep medical history private. Therefore, we provide an open-source implementation of this work.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Priyam Basu (4 papers)
  2. Tiasa Singha Roy (5 papers)
  3. Rakshit Naidu (17 papers)
  4. Zumrut Muftuoglu (2 papers)
  5. Sahib Singh (12 papers)
  6. Fatemehsadat Mireshghallah (26 papers)
Citations (47)

Summary

We haven't generated a summary for this paper yet.