Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Incentivizing Data Contribution in Cross-Silo Federated Learning (2203.03885v2)

Published 8 Mar 2022 in cs.GT

Abstract: In cross-silo federated learning, clients (e.g., organizations) train a shared global model using local data. However, due to privacy concerns, the clients may not contribute enough data points during training. To address this issue, we propose a general incentive framework where the profit/benefit obtained from the global model can be appropriately allocated to clients to incentivize data contribution. We formulate the clients' interactions as a data contribution game and study its equilibrium. We characterize conditions for an equilibrium to exist, and prove that each client's equilibrium data contribution increases in its data quality and decreases in the privacy sensitivity. We further conduct experiments using CIFAR-10 and show that the results are consistent with the analysis. Moreover, we show that practical allocation mechanisms such as linearly proportional, leave-one-out, and Shapley-value incentivize more data contribution from clients with higher-quality data, in which leave-one-out tends to achieve the highest global model accuracy at equilibrium.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Chao Huang (244 papers)
  2. Shuqi Ke (5 papers)
  3. Charles Kamhoua (24 papers)
  4. Prasant Mohapatra (44 papers)
  5. Xin Liu (820 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.