Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Accuracy Improvement in Differentially Private Logistic Regression: A Pre-training Approach (2307.13771v3)

Published 25 Jul 2023 in cs.LG and cs.CR

Abstract: Machine learning (ML) models can memorize training datasets. As a result, training ML models over private datasets can lead to the violation of individuals' privacy. Differential privacy (DP) is a rigorous privacy notion to preserve the privacy of underlying training datasets. Yet, training ML models in a DP framework usually degrades the accuracy of ML models. This paper aims to boost the accuracy of a DP logistic regression (LR) via a pre-training module. In more detail, we initially pre-train our LR model on a public training dataset that there is no privacy concern about it. Then, we fine-tune our DP-LR model with the private dataset. In the numerical results, we show that adding a pre-training module significantly improves the accuracy of the DP-LR model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Mohammad Hoseinpour (1 paper)
  2. Milad Hoseinpour (2 papers)
  3. Ali Aghagolzadeh (5 papers)

Summary

We haven't generated a summary for this paper yet.