2,247 research outputs found

    Performance Evaluation of Caching Policies in NDN - an ICN Architecture

    Full text link
    Information Centric Networking (ICN) advocates the philosophy of accessing the content independent of its location. Owing to this location independence in ICN, the routers en-route can be enabled to cache the content to serve the future requests for the same content locally. Several ICN architectures have been proposed in the literature along with various caching algorithms for caching and cache replacement at the routers en-route. The aim of this paper is to critically evaluate various caching policies using Named Data Networking (NDN), an ICN architecture proposed in literature. We have presented the performance comparison of different caching policies naming First In First Out (FIFO), Least Recently Used (LRU), and Universal Caching (UC) in two network models; Watts-Strogatz (WS) model (suitable for dense short link networks such as sensor networks) and Sprint topology (better suited for large Internet Service Provider (ISP) networks) using ndnSIM, an ns3 based discrete event simulator for NDN architecture. Our results indicate that UC outperforms other caching policies such as LRU and FIFO and makes UC a better alternative for both sensor networks and ISP networks

    Novel applications and contexts for the cognitive packet network

    Get PDF
    Autonomic communication, which is the development of self-configuring, self-adapting, self-optimising and self-healing communication systems, has gained much attention in the network research community. This can be explained by the increasing demand for more sophisticated networking technologies with physical realities that possess computation capabilities and can operate successfully with minimum human intervention. Such systems are driving innovative applications and services that improve the quality of life of citizens both socially and economically. Furthermore, autonomic communication, because of its decentralised approach to communication, is also being explored by the research community as an alternative to centralised control infrastructures for efficient management of large networks. This thesis studies one of the successful contributions in the autonomic communication research, the Cognitive Packet Network (CPN). CPN is a highly scalable adaptive routing protocol that allows for decentralised control in communication. Consequently, CPN has achieved significant successes, and because of the direction of research, we expect it to continue to find relevance. To investigate this hypothesis, we research new applications and contexts for CPN. This thesis first studies Information-Centric Networking (ICN), a future Internet architecture proposal. ICN adopts a data-centric approach such that contents are directly addressable at the network level and in-network caching is easily supported. An optimal caching strategy for an information-centric network is first analysed, and approximate solutions are developed and evaluated. Furthermore, a CPN inspired forwarding strategy for directing requests in such a way that exploits the in-network caching capability of ICN is proposed. The proposed strategy is evaluated via discrete event simulations and shown to be more effective in its search for local cache hits compared to the conventional methods. Finally, CPN is proposed to implement the routing system of an Emergency Cyber-Physical System for guiding evacuees in confined spaces in emergency situations. By exploiting CPN’s QoS capabilities, different paths are assigned to evacuees based on their ongoing health conditions using well-defined path metrics. The proposed system is evaluated via discrete-event simulations and shown to improve survival chances compared to a static system that treats evacuees in the same way.Open Acces

    Content Delivery Latency of Caching Strategies for Information-Centric IoT

    Full text link
    In-network caching is a central aspect of Information-Centric Networking (ICN). It enables the rapid distribution of content across the network, alleviating strain on content producers and reducing content delivery latencies. ICN has emerged as a promising candidate for use in the Internet of Things (IoT). However, IoT devices operate under severe constraints, most notably limited memory. This means that nodes cannot indiscriminately cache all content; instead, there is a need for a caching strategy that decides what content to cache. Furthermore, many applications in the IoT space are timesensitive; therefore, finding a caching strategy that minimises the latency between content request and delivery is desirable. In this paper, we evaluate a number of ICN caching strategies in regards to latency and hop count reduction using IoT devices in a physical testbed. We find that the topology of the network, and thus the routing algorithm used to generate forwarding information, has a significant impact on the performance of a given caching strategy. To the best of our knowledge, this is the first study that focuses on latency effects in ICN-IoT caching while using real IoT hardware, and the first to explicitly discuss the link between routing algorithm, network topology, and caching effects.Comment: 10 pages, 9 figures, journal pape

    Proactive edge caching in content-centric networks with massive dynamic content requests

    Get PDF
    Edge computing is a promising infrastructure evolution to reduce traffic loads and support low-latency communications. Furthermore, content-centric networks provide a natural solution to cache contents at edge nodes. However, it is a challenge for edge nodes to handle massive and highly dynamic content requests by users, and if without an efficient content caching strategy, the edge nodes will encounter high traffic load and latency due to increasing retrieval from content providers. This paper formulates a proactive edge caching problem to minimize the content retrieval cost at edge nodes. We exploit the inherent content caching and request aggregation mechanism in the content-centric networks to jointly minimize traffic load and content retrieval delay cost generated by the massive and dynamic content requests. We develop a Q-learning algorithm, which is an online optimal caching strategy, as it is adaptable to dynamic content popularity and content request intensity, and derive the long-term minimization of the content retrieval cost. Simulation results illustrate that the proposed algorithm can achieve a lower content retrieval cost compared with several baseline caching schemes

    Adaptive Prioritized Probabilistic Caching Algorithm for Content Centric Networks

    Get PDF
    This paper presents an adaptive prioritized probabilistic caching algorithm (APP) for content centric networks (CCN). The objective of the new caching algorithm is to satisfy content requesters with both improving received data quality and maintaining overall network performance. APP allows CCN routers to cache data packets based on the caching probability which is prioritized and unequally handles incoming data packets according to data priorities. APP adjusts the caching probability based on cache events occurred at the CCN router, and the current caching probability is calculated from the previous caching probability. We evaluate APP performance via computer simulations and compare the performance of our caching algorithm with previous caching schemes. The performance evaluation metrics compose of the received data quality, cache-hit percentage, server load, and traffic load. The computer simulation results show that APP yields the better data quality to content requesters and nearly performs in-network caching as well as the previous probabilistic caching scheme

    A Light-Weight Forwarding Plane for Content-Centric Networks

    Full text link
    We present CCN-DART, a more efficient forwarding approach for content-centric networking (CCN) than named data networking (NDN) that substitutes Pending Interest Tables (PIT) with Data Answer Routing Tables (DART) and uses a novel approach to eliminate forwarding loops. The forwarding state required at each router using CCN-DART consists of segments of the routes between consumers and content providers that traverse a content router, rather than the Interests that the router forwards towards content providers. Accordingly, the size of a DART is proportional to the number of routes used by Interests traversing a router, rather than the number of Interests traversing a router. We show that CCN-DART avoids forwarding loops by comparing distances to name prefixes reported by neighbors, even when routing loops exist. Results of simulation experiments comparing CCN-DART with NDN using the ndnSIM simulation tool show that CCN-DART incurs 10 to 20 times less storage overhead
    • …
    corecore