Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deploying Federated Learning in Large-Scale Cellular Networks: Spatial Convergence Analysis (2103.06056v1)

Published 10 Mar 2021 in cs.IT and math.IT

Abstract: The deployment of federated learning in a wireless network, called federated edge learning (FEEL), exploits low-latency access to distributed mobile data to efficiently train an AI model while preserving data privacy. In this work, we study the spatial (i.e., spatially averaged) learning performance of FEEL deployed in a large-scale cellular network with spatially random distributed devices. Both the schemes of digital and analog transmission are considered, providing support of error-free uploading and over-the-air aggregation of local model updates by devices. The derived spatial convergence rate for digital transmission is found to be constrained by a limited number of active devices regardless of device density and converges to the ground-true rate exponentially fast as the number grows. The population of active devices depends on network parameters such as processing gain and signal-to-interference threshold for decoding. On the other hand, the limit does not exist for uncoded analog transmission. In this case, the spatial convergence rate is slowed down due to the direct exposure of signals to the perturbation of inter-cell interference. Nevertheless, the effect diminishes when devices are dense as interference is averaged out by aggressive over-the-air aggregation. In terms of learning latency (in second), analog transmission is preferred to the digital scheme as the former dramatically reduces multi-access latency by enabling simultaneous access.

Citations (36)

Summary

We haven't generated a summary for this paper yet.