2,114 research outputs found
A Content-based Centrality Metric for Collaborative Caching in Information-Centric Fogs
Information-Centric Fog Computing enables a multitude of nodes near the
end-users to provide storage, communication, and computing, rather than in the
cloud. In a fog network, nodes connect with each other directly to get content
locally whenever possible. As the topology of the network directly influences
the nodes' connectivity, there has been some work to compute the graph
centrality of each node within that network topology. The centrality is then
used to distinguish nodes in the fog network, or to prioritize some nodes over
others to participate in the caching fog. We argue that, for an
Information-Centric Fog Computing approach, graph centrality is not an
appropriate metric. Indeed, a node with low connectivity that caches a lot of
content may provide a very valuable role in the network.
To capture this, we introduce acontent-based centrality (CBC) metric which
takes into account how well a node is connected to the content the network is
delivering, rather than to the other nodes in the network. To illustrate the
validity of considering content-based centrality, we use this new metric for a
collaborative caching algorithm. We compare the performance of the proposed
collaborative caching with typical centrality based, non-centrality based, and
non-collaborative caching mechanisms. Our simulation implements CBC on three
instances of large scale realistic network topology comprising 2,896 nodes with
three content replication levels. Results shows that CBC outperforms benchmark
caching schemes and yields a roughly 3x improvement for the average cache hit
rate
Performance Evaluation of Caching Policies in NDN - an ICN Architecture
Information Centric Networking (ICN) advocates the philosophy of accessing
the content independent of its location. Owing to this location independence in
ICN, the routers en-route can be enabled to cache the content to serve the
future requests for the same content locally. Several ICN architectures have
been proposed in the literature along with various caching algorithms for
caching and cache replacement at the routers en-route. The aim of this paper is
to critically evaluate various caching policies using Named Data Networking
(NDN), an ICN architecture proposed in literature. We have presented the
performance comparison of different caching policies naming First In First Out
(FIFO), Least Recently Used (LRU), and Universal Caching (UC) in two network
models; Watts-Strogatz (WS) model (suitable for dense short link networks such
as sensor networks) and Sprint topology (better suited for large Internet
Service Provider (ISP) networks) using ndnSIM, an ns3 based discrete event
simulator for NDN architecture. Our results indicate that UC outperforms other
caching policies such as LRU and FIFO and makes UC a better alternative for
both sensor networks and ISP networks
Offloading Content with Self-organizing Mobile Fogs
Mobile users in an urban environment access content on the internet from
different locations. It is challenging for the current service providers to
cope with the increasing content demand from a large number of collocated
mobile users. In-network caching to offload content at nodes closer to users
alleviate the issue, though efficient cache management is required to find out
who should cache what, when and where in an urban environment, given nodes
limited computing, communication and caching resources. To address this, we
first define a novel relation between content popularity and availability in
the network and investigate a node's eligibility to cache content based on its
urban reachability. We then allow nodes to self-organize into mobile fogs to
increase the distributed cache and maximize content availability in a
cost-effective manner. However, to cater rational nodes, we propose a coalition
game for the nodes to offer a maximum "virtual cache" assuming a monetary
reward is paid to them by the service/content provider. Nodes are allowed to
merge into different spatio-temporal coalitions in order to increase the
distributed cache size at the network edge. Results obtained through
simulations using realistic urban mobility trace validate the performance of
our caching system showing a ratio of 60-85% of cache hits compared to the
30-40% obtained by the existing schemes and 10% in case of no coalition
Cooperative announcement-based caching for video-on-demand streaming
Recently, video-on-demand (VoD) streaming services like Netflix and Hulu have gained a lot of popularity. This has led to a strong increase in bandwidth capacity requirements in the network. To reduce this network load, the design of appropriate caching strategies is of utmost importance. Based on the fact that, typically, a video stream is temporally segmented into smaller chunks that can be accessed and decoded independently, cache replacement strategies have been developed that take advantage of this temporal structure in the video. In this paper, two caching strategies are proposed that additionally take advantage of the phenomenon of binge watching, where users stream multiple consecutive episodes of the same series, reported by recent user behavior studies to become the everyday behavior. Taking into account this information allows us to predict future segment requests, even before the video playout has started. Two strategies are proposed, both with a different level of coordination between the caches in the network. Using a VoD request trace based on binge watching user characteristics, the presented algorithms have been thoroughly evaluated in multiple network topologies with different characteristics, showing their general applicability. It was shown that in a realistic scenario, the proposed election-based caching strategy can outperform the state-of-the-art by 20% in terms of cache hit ratio while using 4% less network bandwidth
- …