274 research outputs found
Cache-Enabled Broadcast Packet Erasure Channels with State Feedback
We consider a cache-enabled K-user broadcast erasure packet channel in which
a server with a library of N files wishes to deliver a requested file to each
user who is equipped with a cache of a finite memory M. Assuming that the
transmitter has state feedback and user caches can be filled during off-peak
hours reliably by decentralized cache placement, we characterize the optimal
rate region as a function of the memory size, the erasure probability. The
proposed delivery scheme, based on the scheme proposed by Gatzianas et al.,
exploits the receiver side information established during the placement phase.
Our results enable us to quantify the net benefits of decentralized coded
caching in the presence of erasure. The role of state feedback is found useful
especially when the erasure probability is large and/or the normalized memory
size is small.Comment: 8 pages, 4 figures, to be presented at the 53rd Annual Allerton
Conference on Communication, Control, and Computing, IL, US
Content Delivery in Erasure Broadcast Channels with Cache and Feedback
We study a content delivery problem in a K-user erasure broadcast channel
such that a content providing server wishes to deliver requested files to
users, each equipped with a cache of a finite memory. Assuming that the
transmitter has state feedback and user caches can be filled during off-peak
hours reliably by the decentralized content placement, we characterize the
achievable rate region as a function of the memory sizes and the erasure
probabilities. The proposed delivery scheme, based on the broadcasting scheme
by Wang and Gatzianas et al., exploits the receiver side information
established during the placement phase. Our results can be extended to
centralized content placement as well as multi-antenna broadcast channels with
state feedback.Comment: 29 pages, 7 figures. A short version has been submitted to ISIT 201
Fundamental Limits of Caching in Wireless D2D Networks
We consider a wireless Device-to-Device (D2D) network where communication is
restricted to be single-hop. Users make arbitrary requests from a finite
library of files and have pre-cached information on their devices, subject to a
per-node storage capacity constraint. A similar problem has already been
considered in an ``infrastructure'' setting, where all users receive a common
multicast (coded) message from a single omniscient server (e.g., a base station
having all the files in the library) through a shared bottleneck link. In this
work, we consider a D2D ``infrastructure-less'' version of the problem. We
propose a caching strategy based on deterministic assignment of subpackets of
the library files, and a coded delivery strategy where the users send linearly
coded messages to each other in order to collectively satisfy their demands. We
also consider a random caching strategy, which is more suitable to a fully
decentralized implementation. Under certain conditions, both approaches can
achieve the information theoretic outer bound within a constant multiplicative
factor. In our previous work, we showed that a caching D2D wireless network
with one-hop communication, random caching, and uncoded delivery, achieves the
same throughput scaling law of the infrastructure-based coded multicasting
scheme, in the regime of large number of users and files in the library. This
shows that the spatial reuse gain of the D2D network is order-equivalent to the
coded multicasting gain of single base station transmission. It is therefore
natural to ask whether these two gains are cumulative, i.e.,if a D2D network
with both local communication (spatial reuse) and coded multicasting can
provide an improved scaling law. Somewhat counterintuitively, we show that
these gains do not cumulate (in terms of throughput scaling law).Comment: 45 pages, 5 figures, Submitted to IEEE Transactions on Information
Theory, This is the extended version of the conference (ITW) paper
arXiv:1304.585
- …