Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adversarial Predictions of Data Distributions Across Federated Internet-of-Things Devices (2308.14658v1)

Published 28 Aug 2023 in cs.LG and cs.DC

Abstract: Federated learning (FL) is increasingly becoming the default approach for training machine learning models across decentralized Internet-of-Things (IoT) devices. A key advantage of FL is that no raw data are communicated across the network, providing an immediate layer of privacy. Despite this, recent works have demonstrated that data reconstruction can be done with the locally trained model updates which are communicated across the network. However, many of these works have limitations with regard to how the gradients are computed in backpropagation. In this work, we demonstrate that the model weights shared in FL can expose revealing information about the local data distributions of IoT devices. This leakage could expose sensitive information to malicious actors in a distributed system. We further discuss results which show that injecting noise into model weights is ineffective at preventing data leakage without seriously harming the global model accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Samir Rajani (2 papers)
  2. Dario Dematties (2 papers)
  3. Nathaniel Hudson (16 papers)
  4. Kyle Chard (87 papers)
  5. Nicola Ferrier (12 papers)
  6. Rajesh Sankaran (2 papers)
  7. Peter Beckman (1 paper)