6,293 research outputs found
Performance Evaluation of Caching Policies in NDN - an ICN Architecture
Information Centric Networking (ICN) advocates the philosophy of accessing
the content independent of its location. Owing to this location independence in
ICN, the routers en-route can be enabled to cache the content to serve the
future requests for the same content locally. Several ICN architectures have
been proposed in the literature along with various caching algorithms for
caching and cache replacement at the routers en-route. The aim of this paper is
to critically evaluate various caching policies using Named Data Networking
(NDN), an ICN architecture proposed in literature. We have presented the
performance comparison of different caching policies naming First In First Out
(FIFO), Least Recently Used (LRU), and Universal Caching (UC) in two network
models; Watts-Strogatz (WS) model (suitable for dense short link networks such
as sensor networks) and Sprint topology (better suited for large Internet
Service Provider (ISP) networks) using ndnSIM, an ns3 based discrete event
simulator for NDN architecture. Our results indicate that UC outperforms other
caching policies such as LRU and FIFO and makes UC a better alternative for
both sensor networks and ISP networks
Content Delivery Latency of Caching Strategies for Information-Centric IoT
In-network caching is a central aspect of Information-Centric Networking
(ICN). It enables the rapid distribution of content across the network,
alleviating strain on content producers and reducing content delivery
latencies. ICN has emerged as a promising candidate for use in the Internet of
Things (IoT). However, IoT devices operate under severe constraints, most
notably limited memory. This means that nodes cannot indiscriminately cache all
content; instead, there is a need for a caching strategy that decides what
content to cache. Furthermore, many applications in the IoT space are
timesensitive; therefore, finding a caching strategy that minimises the latency
between content request and delivery is desirable. In this paper, we evaluate a
number of ICN caching strategies in regards to latency and hop count reduction
using IoT devices in a physical testbed. We find that the topology of the
network, and thus the routing algorithm used to generate forwarding
information, has a significant impact on the performance of a given caching
strategy. To the best of our knowledge, this is the first study that focuses on
latency effects in ICN-IoT caching while using real IoT hardware, and the first
to explicitly discuss the link between routing algorithm, network topology, and
caching effects.Comment: 10 pages, 9 figures, journal pape
A Content-based Centrality Metric for Collaborative Caching in Information-Centric Fogs
Information-Centric Fog Computing enables a multitude of nodes near the
end-users to provide storage, communication, and computing, rather than in the
cloud. In a fog network, nodes connect with each other directly to get content
locally whenever possible. As the topology of the network directly influences
the nodes' connectivity, there has been some work to compute the graph
centrality of each node within that network topology. The centrality is then
used to distinguish nodes in the fog network, or to prioritize some nodes over
others to participate in the caching fog. We argue that, for an
Information-Centric Fog Computing approach, graph centrality is not an
appropriate metric. Indeed, a node with low connectivity that caches a lot of
content may provide a very valuable role in the network.
To capture this, we introduce acontent-based centrality (CBC) metric which
takes into account how well a node is connected to the content the network is
delivering, rather than to the other nodes in the network. To illustrate the
validity of considering content-based centrality, we use this new metric for a
collaborative caching algorithm. We compare the performance of the proposed
collaborative caching with typical centrality based, non-centrality based, and
non-collaborative caching mechanisms. Our simulation implements CBC on three
instances of large scale realistic network topology comprising 2,896 nodes with
three content replication levels. Results shows that CBC outperforms benchmark
caching schemes and yields a roughly 3x improvement for the average cache hit
rate
- …