640 research outputs found

    Revisiting Resource Pooling: The Case for In-Network Resource Sharing.

    Get PDF
    We question the widely adopted view of in-network caches acting as temporary storage for the most popular content in Information-Centric Networks (ICN). Instead, we propose that in-network storage is used as a place of temporary custody for incoming content in a store and forward manner. Given this functionality of in-network storage, senders push content into the network in an open-loop manner to take advantage of underutilised links. When content hits the bottleneck link it gets re-routed through alternative uncongested paths. If alternative paths do not exist, incoming content is temporarily stored in in-network caches, while the system enters a closed-loop, back-pressure mode of operation to avoid congestive collapse. Our proposal follows in spirit the resource pooling principle, which, however, is restricted to end-to-end resources and paths. We extend this principle to also take advantage of in-network resources, in terms of multiplicity of available sub-paths (as compared to multihomed users only) and in-network cache space. We call the proposed principle In-Network Resource Pooling Principle (INRPP). Using the INRPP, congestion, or increased contention over a link, is dealt with locally in a hop-by-hop manner, instead of end-to-end. INRPP utilises resources throughout the network more efficiently and opens up new directions for research in the multipath routing and congestion control areas

    Mind the gap: modelling video delivery under expected periods of disconnection.

    Get PDF
    In this work we model video delivery under expected periods of disconnection, such as the ones experienced in public transportation systems. Our main goal is to quantify the gains of users' collaboration in terms of Quality of Experience (QoE) in the context of intermittently available and bandwidth-limited WiFi connectivity. Under the assumption that Wi-Fi connectivity is available within underground stations, but absent between them, at first, we define a mathematical model which describes the content distribution under these conditions and we present the users' QoE function in terms of undisrupted video playback. Next, we expand this model to include the case of collaboration between users for content sharing in a peer-to-peer (P2P) way. Lastly, we evaluate our model based on real data from the London Underground network, where we investigate the feasibility of content distribution, only to find that collaboration between users increases significantly their QoE

    NFaaS: Named Function as a Service

    Get PDF
    In the past, the Information-centric networking (ICN) community has focused on issues mainly pertaining to traditional content delivery (e.g., routing and forwarding scalability, congestion control and in-network caching). However, to keep up with future Internet architectural trends the wider area of future Internet paradigms, there is a pressing need to support edge/fog computing environments, where cloud functionality is available more proximate to where the data is generated and needs processing. With this goal in mind, we propose Named Function as a Service (NFaaS), a framework that extends the Named Data Networking architecture to support in-network function execution. In contrast to existing works, NFaaSbuilds on very lightweight VMs and allows for dynamic execution of custom code. Functions can be downloaded and run by any node in the network. Functions can move between nodes according to user demand, making resolution of moving functions a first-class challenge. NFaaSincludes a Kernel Store component, which is responsible not only for storing functions, but also for making decisions on which functions to run locally. NFaaSincludes a routing protocol and a number of forwarding strategies to deploy and dynamically migrate functions within the network. We validate our design through extensive simulations, which show that delay-sensitive functions are deployed closer to the edge, while less delay-sensitive ones closer to the core

    Proof-of-Prestige: A Useful Work Reward System for Unverifiable Tasks

    Full text link
    As cryptographic tokens and altcoins are increasingly being built to serve as utility tokens, the notion of useful work consensus protocols, as opposed to number-crunching PoW consensus, is becoming ever more important. In such contexts, users get rewards from the network after they have carried out some specific task useful for the network. While in some cases the proof of some utility or service can be proved, the majority of tasks are impossible to verify. In order to deal with such cases, we design Proof-of-Prestige (PoP) - a reward system that can run on top of Proof-of-Stake blockchains. PoP introduces prestige which is a volatile resource and, in contrast to coins, regenerates over time. Prestige can be gained by performing useful work, spent when benefiting from services and directly translates to users minting power. PoP is resistant against Sybil and Collude attacks and can be used to reward workers for completing unverifiable tasks, while keeping the system free for the end-users. We use two exemplar use-cases to showcase the usefulness of PoP and we build a simulator to assess the cryptoeconomic behaviour of the system in terms of prestige transfer between nodes.Comment: 2019 IEEE International Conference on Blockchain and Cryptocurrency (ICBC 2019

    On Energy Reduction and Green Networking Enhancement due to In-Network Caching

    Get PDF
    In-network caching in information centric networking (ICN) is considered as a promising approach to reducing energy consumption of an entire network. However, it is also considered as an energy consuming technique. These contradictory claims lead to one research question: Does caching really reduce the energy consumption of the entire network? To answer the question, we formulate an ICN network as an optimization problem with a realistic energy consumption model for an ICN router. By solving the formulation assuming that ICN forwarding software currently under development is used as a forwarding engine of an ICN router, we reveal that in-network caching alone does not reduce much energy but it enhances a currently developed green networking technique even though the forwarding engine is not fully optimized

    Efficient Hash-routing and Domain Clustering Techniques for Information-Centric Networks

    Get PDF
    Hash-routing is a well-known technique used in server-cluster environments to direct content requests to the responsible servers hosting the requested content. In this work, we look at hash-routing from a different angle and apply the technique to Information-Centric Networking (ICN) environments, where in-network content caches serve as temporary storage for content. In particular, edge-domain routers re-direct requests to in-network caches, more often than not off the shortest path, according to the hash-assignment function. Although the benefits of this off-path in-network caching scheme are significant (e.g., high cache hit rate with minimal co-ordination overhead), the basic scheme comes with disadvantages. That is, in case of very large domains the off-path detour of requests might increase latency to prohibitive levels. In order to deal with extensive detour delays, we investigate nodal/domain clustering techniques, according to which large domains are split in clusters, which in turn apply hash-routing in the subset of nodes of each cluster. We model and evaluate the behaviour of nodal clustering and report significant improvement in delivery latency, which comes at the cost of a slight decrease in cache hit rates (i.e., up to 50% improvement in delivery latency for less than 10% decrease in cache hit rate compared to the original hash-routing scheme applied in the whole domain)

    A native content discovery mechanism for the information-centric networks

    Get PDF
    Recent research has considered various approaches for discovering content in the cache-enabled nodes of an Autonomous System (AS) to reduce the costly inter-AS traffic. Such approaches include i) searching content opportunistically (on-path) along the default intra-AS path towards the content origin for limited gain, and ii) actively coordinate nodes when caching content for significantly higher gains, but also higher overhead. In this paper, we try to combine the merits of both worlds by using traditional opportunistic caching mechanisms enhanced with a lightweight content discovery approach. Particularly, a content retrieved through an inter-AS link is cached only once along the intra-AS delivery path to maximize network storage utilization, and ephemeral forwarding state to locate temporarily stored content is established opportunistically at each node along that path during the processing of Data packets. The ephemeral forwarding state either points to the arriving or the destination face of the Data packet depending on whether the content has already been cached along the path or not. The challenge in such an approach is to appropriately use and maintain the ephemeral forwarding state to minimize inter-AS content retrieval, while keeping retrieval latency and overhead at acceptable levels. We propose several forwarding strategies to use and manage ephemeral state and evaluate our mechanism using an ISP topology for various system parameters. Our results indicate that our opportunistic content discovery mechanism can achieve near-optimal performance and significantly reduce inter-AS traffic
    corecore