Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Wyner's Network on Caches: Combining Receiver Caching with a Flexible Backhaul (1903.06515v1)

Published 15 Mar 2019 in cs.IT and math.IT

Abstract: In this work, we study a large linear interference network with an equal number of transmitters and receivers, where each transmitter is connected to two subsequent receivers. Each transmitter has individual access to a backhaul link (fetching the equivalent of $M_{T}$ files), while each receiver can cache a fraction $\gamma$ of the library. We explore the tradeoff between the communication rate, backhaul load, and caching storage by designing algorithms that can harness the benefits of cooperative transmission in partially connected networks, while exploiting the advantages of multicast transmissions attributed to user caching. We show that receiver caching and fetching content from the backhaul are two resources that can simultaneously increase the delivery performance in synergistic ways. Specifically, an interesting outcome of this work is that user caching of a fraction $\gamma$ of the library can increase the per-user Degrees of Freedom (puDoF) by $\gamma$. Further, the results reveal significant savings in the backhaul load, even in the small cache size region. For example, the puDoF achieved using the pair $(M_{T}=8, \gamma=0)$ can also be achieved with the pairs $(M_{T}=4,\gamma=0.035)$ and $(M_{T}=2,\gamma=0.1)$, showing that even small caches can provide significant savings in the backhaul load.

Citations (2)

Summary

We haven't generated a summary for this paper yet.