970 research outputs found
Caching with Partial Adaptive Matching
We study the caching problem when we are allowed to match each user to one of
a subset of caches after its request is revealed. We focus on non-uniformly
popular content, specifically when the file popularities obey a Zipf
distribution. We study two extremal schemes, one focusing on coded server
transmissions while ignoring matching capabilities, and the other focusing on
adaptive matching while ignoring potential coding opportunities. We derive the
rates achieved by these schemes and characterize the regimes in which one
outperforms the other. We also compare them to information-theoretic outer
bounds, and finally propose a hybrid scheme that generalizes ideas from the two
schemes and performs at least as well as either of them in most memory regimes.Comment: 35 pages, 7 figures. Shorter versions have appeared in IEEE ISIT 2017
and IEEE ITW 201
Effect of Number of Users in Multi-level Coded Caching
It has been recently established that joint design of content delivery and
storage (coded caching) can significantly improve performance over conventional
caching. This has also been extended to the case when content has non-uniform
popularity through several models. In this paper we focus on a multi-level
popularity model, where content is divided into levels based on popularity. We
consider two extreme cases of user distribution across caches for the
multi-level popularity model: a single user per cache (single-user setup)
versus a large number of users per cache (multi-user setup). When the capacity
approximation is universal (independent of number of popularity levels as well
as number of users, files and caches), we demonstrate a dichotomy in the
order-optimal strategies for these two extreme cases. In the multi-user case,
sharing memory among the levels is order-optimal, whereas for the single-user
case clustering popularity levels and allocating all the memory to them is the
order-optimal scheme. In proving these results, we develop new
information-theoretic lower bounds for the problem.Comment: 13 pages; 2 figures. A shorter version is to appear in IEEE ISIT 201
An Efficient Coded Multicasting Scheme Preserving the Multiplicative Caching Gain
Coded multicasting has been shown to be a promis- ing approach to
significantly improve the caching performance of content delivery networks with
multiple caches downstream of a common multicast link. However, achievable
schemes proposed to date have been shown to achieve the proved order-optimal
performance only in the asymptotic regime in which the number of packets per
requested item goes to infinity. In this paper, we first extend the asymptotic
analysis of the achievable scheme in [1], [2] to the case of heterogeneous
cache sizes and demand distributions, providing the best known upper bound on
the fundamental limiting performance when the number of packets goes to
infinity. We then show that the scheme achieving this upper bound quickly loses
its multiplicative caching gain for finite content packetization. To overcome
this limitation, we design a novel polynomial-time algorithm based on random
greedy graph- coloring that, while keeping the same finite content
packetization, recovers a significant part of the multiplicative caching gain.
Our results show that the order-optimal coded multicasting schemes proposed to
date, while useful in quantifying the fundamental limiting performance, must be
properly designed for practical regimes of finite packetization.Comment: 6 pages, 7 figures, Published in Infocom CNTCV 201
A Learning-Based Approach to Caching in Heterogenous Small Cell Networks
A heterogenous network with base stations (BSs), small base stations (SBSs)
and users distributed according to independent Poisson point processes is
considered. SBS nodes are assumed to possess high storage capacity and to form
a distributed caching network. Popular files are stored in local caches of
SBSs, so that a user can download the desired files from one of the SBSs in its
vicinity. The offloading-loss is captured via a cost function that depends on
the random caching strategy proposed here. The popularity profile of cached
content is unknown and estimated using instantaneous demands from users within
a specified time interval. An estimate of the cost function is obtained from
which an optimal random caching strategy is devised. The training time to
achieve an difference between the achieved and optimal costs is
finite provided the user density is greater than a predefined threshold, and
scales as , where is the support of the popularity profile. A transfer
learning-based approach to improve this estimate is proposed. The training time
is reduced when the popularity profile is modeled using a parametric family of
distributions; the delay is independent of and scales linearly with the
dimension of the distribution parameter.Comment: 12 pages, 5 figures, published in IEEE Transactions on
Communications, 2016. arXiv admin note: text overlap with arXiv:1504.0363
Adaptive Delivery in Caching Networks
The problem of content delivery in caching networks is investigated for
scenarios where multiple users request identical files. Redundant user demands
are likely when the file popularity distribution is highly non-uniform or the
user demands are positively correlated. An adaptive method is proposed for the
delivery of redundant demands in caching networks. Based on the redundancy
pattern in the current demand vector, the proposed method decides between the
transmission of uncoded messages or the coded messages of [1] for delivery.
Moreover, a lower bound on the delivery rate of redundant requests is derived
based on a cutset bound argument. The performance of the adaptive method is
investigated through numerical examples of the delivery rate of several
specific demand vectors as well as the average delivery rate of a caching
network with correlated requests. The adaptive method is shown to considerably
reduce the gap between the non-adaptive delivery rate and the lower bound. In
some specific cases, using the adaptive method, this gap shrinks by almost 50%
for the average rate.Comment: 8 pages,8 figures. Submitted to IEEE transaction on Communications in
2015. A short version of this article was published as an IEEE Communications
Letter with DOI: 10.1109/LCOMM.2016.255814
- …