3,228 research outputs found
A Literature Survey of Cooperative Caching in Content Distribution Networks
Content distribution networks (CDNs) which serve to deliver web objects
(e.g., documents, applications, music and video, etc.) have seen tremendous
growth since its emergence. To minimize the retrieving delay experienced by a
user with a request for a web object, caching strategies are often applied -
contents are replicated at edges of the network which is closer to the user
such that the network distance between the user and the object is reduced. In
this literature survey, evolution of caching is studied. A recent research
paper [15] in the field of large-scale caching for CDN was chosen to be the
anchor paper which serves as a guide to the topic. Research studies after and
relevant to the anchor paper are also analyzed to better evaluate the
statements and results of the anchor paper and more importantly, to obtain an
unbiased view of the large scale collaborate caching systems as a whole.Comment: 5 pages, 5 figure
Cooperative announcement-based caching for video-on-demand streaming
Recently, video-on-demand (VoD) streaming services like Netflix and Hulu have gained a lot of popularity. This has led to a strong increase in bandwidth capacity requirements in the network. To reduce this network load, the design of appropriate caching strategies is of utmost importance. Based on the fact that, typically, a video stream is temporally segmented into smaller chunks that can be accessed and decoded independently, cache replacement strategies have been developed that take advantage of this temporal structure in the video. In this paper, two caching strategies are proposed that additionally take advantage of the phenomenon of binge watching, where users stream multiple consecutive episodes of the same series, reported by recent user behavior studies to become the everyday behavior. Taking into account this information allows us to predict future segment requests, even before the video playout has started. Two strategies are proposed, both with a different level of coordination between the caches in the network. Using a VoD request trace based on binge watching user characteristics, the presented algorithms have been thoroughly evaluated in multiple network topologies with different characteristics, showing their general applicability. It was shown that in a realistic scenario, the proposed election-based caching strategy can outperform the state-of-the-art by 20% in terms of cache hit ratio while using 4% less network bandwidth
Echo State Networks for Proactive Caching in Cloud-Based Radio Access Networks with Mobile Users
In this paper, the problem of proactive caching is studied for cloud radio
access networks (CRANs). In the studied model, the baseband units (BBUs) can
predict the content request distribution and mobility pattern of each user,
determine which content to cache at remote radio heads and BBUs. This problem
is formulated as an optimization problem which jointly incorporates backhaul
and fronthaul loads and content caching. To solve this problem, an algorithm
that combines the machine learning framework of echo state networks with
sublinear algorithms is proposed. Using echo state networks (ESNs), the BBUs
can predict each user's content request distribution and mobility pattern while
having only limited information on the network's and user's state. In order to
predict each user's periodic mobility pattern with minimal complexity, the
memory capacity of the corresponding ESN is derived for a periodic input. This
memory capacity is shown to be able to record the maximum amount of user
information for the proposed ESN model. Then, a sublinear algorithm is proposed
to determine which content to cache while using limited content request
distribution samples. Simulation results using real data from Youku and the
Beijing University of Posts and Telecommunications show that the proposed
approach yields significant gains, in terms of sum effective capacity, that
reach up to 27.8% and 30.7%, respectively, compared to random caching with
clustering and random caching without clustering algorithm.Comment: Accepted in the IEEE Transactions on Wireless Communication
Edge-Caching Wireless Networks: Performance Analysis and Optimization
Edge-caching has received much attention as an efficient technique to reduce
delivery latency and network congestion during peak-traffic times by bringing
data closer to end users. Existing works usually design caching algorithms
separately from physical layer design. In this paper, we analyse edge-caching
wireless networks by taking into account the caching capability when designing
the signal transmission. Particularly, we investigate multi-layer caching where
both base station (BS) and users are capable of storing content data in their
local cache and analyse the performance of edge-caching wireless networks under
two notable uncoded and coded caching strategies. Firstly, we propose a coded
caching strategy that is applied to arbitrary values of cache size. The
required backhaul and access rates are derived as a function of the BS and user
cache size. Secondly, closed-form expressions for the system energy efficiency
(EE) corresponding to the two caching methods are derived. Based on the derived
formulas, the system EE is maximized via precoding vectors design and
optimization while satisfying a predefined user request rate. Thirdly, two
optimization problems are proposed to minimize the content delivery time for
the two caching strategies. Finally, numerical results are presented to verify
the effectiveness of the two caching methods.Comment: to appear in IEEE Trans. Wireless Commu
Flexpop: A popularity-based caching strategy for multimedia applications in information-centric networking
Information-Centric Networking (ICN) is the dominant architecture for the future Internet.
In ICN, the content items are stored temporarily in network nodes such as routers. When the memory of routers becomes full and there is no room for a new
arriving content, the stored contents are evicted to cope with the limited cache size of the routers. Therefore, it is crucial to develop an effective caching strategy for keeping popular contents for a longer period of time. This study proposes a new caching strategy, named Flexible Popularity-based Caching (FlexPop) for storing popular contents.
The FlexPop comprises two mechanisms, i.e., Content Placement Mechanism (CPM), which is responsible for content caching, and Content Eviction Mechanism
(CEM) that deals with content eviction when the router cache is full and there is no space for the new incoming content. Both mechanisms are validated using Fuzzy Set Theory, following the Design Research Methodology (DRM) to manifest that the research is rigorous and repeatable under comparable conditions. The performance of FlexPop is evaluated through simulations and the results are compared with those of the Leave Copy Everywhere (LCE), ProbCache, and Most Popular Content (MPC) strategies. The results show that the FlexPop strategy outperforms LCE, ProbCache,
and MPC with respect to cache hit rate, redundancy, content retrieval delay, memory utilization, and stretch ratio, which are regarded as extremely important metrics (in various studies) for the evaluation of ICN caching. The outcomes exhibited in this study are noteworthy in terms of making FlexPop acceptable to users as they can verify
the performance of ICN before selecting the right caching strategy. Thus FlexPop has potential in the use of ICN for the future Internet such as in deployment of the IoT technology
Exploring the Memory-Bandwidth Tradeoff in an Information-Centric Network
An information-centric network should realize significant economies by
exploiting a favourable memory-bandwidth tradeoff: it is cheaper to store
copies of popular content close to users than to fetch them repeatedly over the
Internet. We evaluate this tradeoff for some simple cache network structures
under realistic assumptions concerning the size of the content catalogue and
its popularity distribution. Derived cost formulas reveal the relative impact
of various cost, traffic and capacity parameters, allowing an appraisal of
possible future network architectures. Our results suggest it probably makes
more sense to envisage the future Internet as a loosely interconnected set of
local data centers than a network like today's with routers augmented by
limited capacity content stores.Comment: Proceedings of ITC 25 (International Teletraffic Congress), Shanghai,
September, 201
- …