772 research outputs found
A Novel Centralized Strategy for Coded Caching with Non-uniform Demands
Despite significant progress in the caching literature concerning the worst
case and uniform average case regimes, the algorithms for caching with
nonuniform demands are still at a basic stage and mostly rely on simple
grouping and memory-sharing techniques. In this work we introduce a novel
centralized caching strategy for caching with nonuniform file popularities. Our
scheme allows for assigning more cache to the files which are more likely to be
requested, while maintaining the same sub-packetization for all the files. As a
result, in the delivery phase it is possible to perform linear codes across
files with different popularities without resorting to zero-padding or
concatenation techniques. We will describe our placement strategy for arbitrary
range of parameters. The delivery phase will be outlined for a small example
for which we are able to show a noticeable improvement over the state of the
art.Comment: 4 pages, 3 figures, submitted to the 2018 International Zurich
Seminar on Information and Communicatio
A Novel Centralized Strategy for Coded Caching with Non-uniform Demands
Despite significant progress in the caching literature concerning the worst case and uniform average case regimes, the algorithms for caching with nonuniform demands are still at a basic stage and mostly rely on simple grouping and memory-sharing techniques. In this work we introduce a novel centralized caching strategy for caching with nonuniform file popularities. Our scheme allows for assigning more cache to the files which are more likely to be requested, while maintaining the same sub-packetization for all the files. As a result, in the delivery phase it is possible to perform linear codes across files with different popularities without resorting to zero-padding or concatenation techniques. We will describe our placement strategy for arbitrary range of parameters. The delivery phase will be outlined for a small example for which we are able to show a noticeable improvement over the state of the art
Fundamental Limits of Caching with Secure Delivery
Caching is emerging as a vital tool for alleviating the severe capacity
crunch in modern content-centric wireless networks. The main idea behind
caching is to store parts of popular content in end-users' memory and leverage
the locally stored content to reduce peak data rates. By jointly designing
content placement and delivery mechanisms, recent works have shown order-wise
reduction in transmission rates in contrast to traditional methods. In this
work, we consider the secure caching problem with the additional goal of
minimizing information leakage to an external wiretapper. The fundamental cache
memory vs. transmission rate trade-off for the secure caching problem is
characterized. Rather surprisingly, these results show that security can be
introduced at a negligible cost, particularly for large number of files and
users. It is also shown that the rate achieved by the proposed caching scheme
with secure delivery is within a constant multiplicative factor from the
information-theoretic optimal rate for almost all parameter values of practical
interest
On Caching with More Users than Files
Caching appears to be an efficient way to reduce peak hour network traffic
congestion by storing some content at the user's cache without knowledge of
later demands. Recently, Maddah-Ali and Niesen proposed a two-phase, placement
and delivery phase, coded caching strategy for centralized systems (where
coordination among users is possible in the placement phase), and for
decentralized systems. This paper investigates the same setup under the further
assumption that the number of users is larger than the number of files. By
using the same uncoded placement strategy of Maddah-Ali and Niesen, a novel
coded delivery strategy is proposed to profit from the multicasting
opportunities that arise because a file may be demanded by multiple users. The
proposed delivery method is proved to be optimal under the constraint of
uncoded placement for centralized systems with two files, moreover it is shown
to outperform known caching strategies for both centralized and decentralized
systems.Comment: 6 pages, 3 figures, submitted to ISIT 201
Uncoded Caching and Cross-level Coded Delivery for Non-uniform File Popularity
Proactive content caching at user devices and coded delivery is studied
considering a non-uniform file popularity distribution. A novel centralized
uncoded caching and coded delivery scheme, which can be applied to large file
libraries, is proposed. The proposed cross-level coded delivery (CLCD) scheme
is shown to achieve a lower average delivery rate than the state of art. In the
proposed CLCD scheme, the same subpacketization is used for all the files in
the library in order to prevent additional zero-padding in the delivery phase,
and unlike the existing schemes in the literature, two users requesting files
from different popularity groups can be served by the same multicast message in
order to reduce the delivery rate. Simulation results indicate significant
reduction in the average delivery rate for typical Zipf distribution parameter
values.Comment: A shorter version of this paper has been presented at IEEE
International Conference on Communications (ICC) 201
- …