56 research outputs found

    Crosslayer Survivability in Overlay-IP-WDM Networks

    Get PDF
    As the Internet moves towards a three-layer architecture consisting of overlay networks on top of the IP network layer on top of WDM-based physical networks, incorporating the interaction between and among network layers is crucial for efficient and effective implementation of survivability.This dissertation has four major foci as follows: First, a first-of-its-kind analysis of the impact of overlay network dependency on the lower layer network unveils that backhaul, a link loop that occurs at any two or more lower layers below the layer where traffic is present, could happen. This prompts our proposal of a crosslayer survivable mapping to highlight such challenges and to offer survivability in an efficient backhaul-free way. The results demonstrate that the impact of layer dependency is more severe than initially anticipated making it clear that independent single layer network design is inadequate to assure service guarantees and efficient capacity allocation. Second, a forbidden link matrix is proposed masking part of the network for use in situations where some physical links are reserved exclusively for a designated service, mainly for the context of providing multiple levels of differentiation on the network use and service guarantee. The masking effect is evaluated on metrics using practical approaches in a sample real-world network, showing that both efficiency and practicality can be achieved. Third, matrix-based optimization problem formulations of several crosslayer survivable mappings are presented; examples on the link availability mapping are particularly illustrated. Fourth, survivability strategies for two-layer backbone networks where traffic originates at each layer are investigated. Optimization-based formulations of performing recovery mechanisms at each layer for both layers of traffic are also presented. Numerical results indicate that, in such a wavelength-based optical network, implementing survivability of all traffic at the bottom layer can be a viable solution with significant advantages.This dissertation concludes by identifying a roadmap of potential future work for crosslayer survivability in layered network settings

    Data Driven Network Design for Cloud Services Based on Historic Utilization

    Get PDF
    In recent years we have seen a shift from traditional networking in enterprises with Data Center centric architectures moving to cloud services. Companies are moving away from private networking technologies like MPLS as they migrate their application workloads to the cloud. With these migrations, network architects must struggle with how to design and build new network infrastructure to support the cloud for all their end users including office workers, remote workers, and home office workers. The main goal for network design is to maximize availability and performance and minimize cost. However, network architects and network engineers tend to over provision networks by sizing the bandwidth for worst case scenarios wasting millions of dollars per year. This thesis will analyze traditional network utilization data from twenty-five of the Fortune 500 companies in the United States and determine the most efficient bandwidth to support cloud services from providers like Amazon, Microsoft, Google, and others. The analysis of real-world data and the resulting proposed scaling factor is an original contribution from this study

    A survey of strategies for communication networks to protect against large-scale natural disasters

    Get PDF
    Recent natural disasters have revealed that emergency networks presently cannot disseminate the necessary disaster information, making it difficult to deploy and coordinate relief operations. These disasters have reinforced the knowledge that telecommunication networks constitute a critical infrastructure of our society, and the urgency in establishing protection mechanisms against disaster-based disruptions

    Optimization Methods for Optical Long-Haul and Access Networks

    Get PDF
    Optical communications based on fiber optics and the associated technologies have seen remarkable progress over the past two decades. Widespread deployment of optical fiber has been witnessed in backbone and metro networks as well as access segments connecting to customer premises and homes. Designing and developing a reliable, robust and efficient end-to-end optical communication system have thus emerged as topics of utmost importance both to researchers and network operators. To fulfill these requirements, various problems have surfaced and received attention, such as network planning, capacity placement, traffic grooming, traffic scheduling, and bandwidth allocation. The optimal network design aims at addressing (one or more of) these problems based on some optimization objectives. In this thesis, we consider two of the most important problems in optical networks; namely the survivability in optical long-haul networks and the problem of bandwidth allocation and scheduling in optical access networks. For the former, we present efficient and accurate models for availability-aware design and service provisioning in p-cycle based survivable networks. We also derive optimization models for survivable network design based on p-trail, a more general protection structure, and compare its performance with p-cycles. Indeed, major cost savings can be obtained when the optical access and long-haul subnetworks become closer to each other by means of consolidation of access and metro networks. As this distance between long-haul and access networks reduces, and the need and expectations from passive optical access networks (PONs) soar, it becomes crucial to efficiently manage bandwidth in the access while providing the desired level of service availability in the long-haul backbone. We therefore address in this thesis the problem of bandwidth management and scheduling in passive optical networks; we design efficient joint and non-joint scheduling and bandwidth allocation methods for multichannel PON as well as next generation 10Gbps Ethernet PON (10G-EPON) while addressing the problem of coexistence between 10G-EPONs and multichannel PONs

    Optimization of traffic flows in multiservice telecomunications networks

    Get PDF
    This dissertation investigates routing optimization in IP telecommunication networks, under normal working conditions as well as under failure conditions. The main objectives of the present optimization procedure are the minimization of the maximum link utilization in the network and to provide a configuration that guarantees a 100% survivability degree. Traditionally two different steps are used to achieve this goal. The first one aims to solve the well known “General Routing Problem (GRP)” in order to find the optimal routing network configuration and, successively, a set of “optimal” backup paths is found in order to guarantee network survivability. Furthermore, traditional survivable techniques assume that the planning tasks are performed in a network control center while restoration schemes are implemented distributively in network nodes. In this dissertation innovative linear programming models are presented that, making use of the Multi Protocol Label Switching – Traffic Engineering (MPLS-TE) techniques and IS-IS/OSPF IP routing protocol, melt routing and survivability requirements. The models are extremely flexible, thus it is possible to improve the objective function in order to fit itself to newer applications and/or traffic typologies. The models presented in this dissertation help Internet Service Providers to optimize their network resources and to guarantee connectivity in case of failure, while still be able to offer a good quality of service

    Modelling and Design of Resilient Networks under Challenges

    Get PDF
    Communication networks, in particular the Internet, face a variety of challenges that can disrupt our daily lives resulting in the loss of human lives and significant financial costs in the worst cases. We define challenges as external events that trigger faults that eventually result in service failures. Understanding these challenges accordingly is essential for improvement of the current networks and for designing Future Internet architectures. This dissertation presents a taxonomy of challenges that can help evaluate design choices for the current and Future Internet. Graph models to analyse critical infrastructures are examined and a multilevel graph model is developed to study interdependencies between different networks. Furthermore, graph-theoretic heuristic optimisation algorithms are developed. These heuristic algorithms add links to increase the resilience of networks in the least costly manner and they are computationally less expensive than an exhaustive search algorithm. The performance of networks under random failures, targeted attacks, and correlated area-based challenges are evaluated by the challenge simulation module that we developed. The GpENI Future Internet testbed is used to conduct experiments to evaluate the performance of the heuristic algorithms developed

    Study, evaluation and contributions to new algorithms for the embedding problem in a network virtualization environment

    Get PDF
    Network virtualization is recognized as an enabling technology for the future Internet. It aims to overcome the resistance of the current Internet to architectural change and to enable a new business model decoupling the network services from the underlying infrastructure. The problem of embedding virtual networks in a substrate network is the main resource allocation challenge in network virtualization and is usually referred to as the Virtual Network Embedding (VNE) problem. VNE deals with the allocation of virtual resources both in nodes and links. Therefore, it can be divided into two sub-problems: Virtual Node Mapping where virtual nodes have to be allocated in physical nodes and Virtual Link Mapping where virtual links connecting these virtual nodes have to be mapped to paths connecting the corresponding nodes in the substrate network. Application of network virtualization relies on algorithms that can instantiate virtualized networks on a substrate infrastructure, optimizing the layout for service-relevant metrics. This class of algorithms is commonly known as VNE algorithms. This thesis proposes a set of contributions to solve the research challenges of the VNE that have not been tackled by the research community. To do that, it performs a deep and comprehensive survey of virtual network embedding. The first research challenge identified is the lack of proposals to solve the virtual link mapping stage of VNE using single path in the physical network. As this problem is NP-hard, existing proposals solve it using well known shortest path algorithms that limit the mapping considering just one constraint. This thesis proposes the use of a mathematical multi-constraint routing framework called paths algebra to solve the virtual link mapping stage. Besides, the thesis introduces a new demand caused by virtual link demands into physical nodes acting as intermediate (hidden) hops in a path of the physical network. Most of the current VNE approaches are centralized. They suffer of scalability issues and provide a single point of failure. In addition, they are not able to embed virtual network requests arriving at the same time in parallel. To solve this challenge, this thesis proposes a distributed, parallel and universal virtual network embedding framework. The proposed framework can be used to run any existing embedding algorithm in a distributed way. Thereby, computational load for embedding multiple virtual networks is spread across the substrate network Energy efficiency is one of the main challenges in future networking environments. Network virtualization can be used to tackle this problem by sharing hardware, instead of requiring dedicated hardware for each instance. Until now, VNE algorithms do not consider energy as a factor for the mapping. This thesis introduces the energy aware VNE where the main objective is to switch off as many network nodes and interfaces as possible by allocating the virtual demands to a consolidated subset of active physical networking equipment. To evaluate and validate the aforementioned VNE proposals, this thesis helped in the development of a software framework called ALgorithms for Embedding VIrtual Networks (ALEVIN). ALEVIN allows to easily implement, evaluate and compare different VNE algorithms according to a set of metrics, which evaluate the algorithms and compute their results on a given scenario for arbitrary parameters

    A framework for traffic flow survivability in wireless networks prone to multiple failures and attacks

    Get PDF
    Transmitting packets over a wireless network has always been challenging due to failures that have always occurred as a result of many types of wireless connectivity issues. These failures have caused significant outages, and the delayed discovery and diagnostic testing of these failures have exacerbated their impact on servicing, economic damage, and social elements such as technological trust. There has been research on wireless network failures, but little on multiple failures such as node-node, node-link, and link–link failures. The problem of capacity efficiency and fast recovery from multiple failures has also not received attention. This research develops a capacity efficient evolutionary swarm survivability framework, which encompasses enhanced genetic algorithm (EGA) and ant colony system (ACS) survivability models to swiftly resolve node-node, node-link, and link-link failures for improved service quality. The capacity efficient models were tested on such failures at different locations on both small and large wireless networks. The proposed models were able to generate optimal alternative paths, the bandwidth required for fast rerouting, minimized transmission delay, and ensured the rerouting path fitness and good transmission time for rerouting voice, video and multimedia messages. Increasing multiple link failures reveal that as failure increases, the bandwidth used for rerouting and transmission time also increases. This implies that, failure increases bandwidth usage which leads to transmission delay, which in turn slows down message rerouting. The suggested framework performs better than the popular Dijkstra algorithm, proactive, adaptive and reactive models, in terms of throughput, packet delivery ratio (PDR), speed of transmission, transmission delay and running time. According to the simulation results, the capacity efficient ACS has a PDR of 0.89, the Dijkstra model has a PDR of 0.86, the reactive model has a PDR of 0.83, the proactive model has a PDR of 0.83, and the adaptive model has a PDR of 0.81. Another performance evaluation was performed to compare the proposed model's running time to that of other evaluated routing models. The capacity efficient ACS model has a running time of 169.89ms on average, while the adaptive model has a running time of 1837ms and Dijkstra has a running time of 280.62ms. With these results, capacity efficient ACS outperforms other evaluated routing algorithms in terms of PDR and running time. According to the mean throughput determined to evaluate the performance of the following routing algorithms: capacity efficient EGA has a mean throughput of 621.6, Dijkstra has a mean throughput of 619.3, proactive (DSDV) has a mean throughput of 555.9, and reactive (AODV) has a mean throughput of 501.0. Since Dijkstra is more similar to proposed models in terms of performance, capacity efficient EGA was compared to Dijkstra as follows: Dijkstra has a running time of 3.8908ms and EGA has a running time of 3.6968ms. In terms of running time and mean throughput, the capacity efficient EGA also outperforms the other evaluated routing algorithms. The generated alternative paths from these investigations demonstrate that the proposed framework works well in preventing the problem of data loss in transit and ameliorating congestion issue resulting from multiple failures and server overload which manifests when the process hangs. The optimal solution paths will in turn improve business activities through quality data communications for wireless service providers.School of ComputingPh. D. (Computer Science

    Autonomous Recovery Of Reconfigurable Logic Devices Using Priority Escalation Of Slack

    Get PDF
    Field Programmable Gate Array (FPGA) devices offer a suitable platform for survivable hardware architectures in mission-critical systems. In this dissertation, active dynamic redundancy-based fault-handling techniques are proposed which exploit the dynamic partial reconfiguration capability of SRAM-based FPGAs. Self-adaptation is realized by employing reconfiguration in detection, diagnosis, and recovery phases. To extend these concepts to semiconductor aging and process variation in the deep submicron era, resilient adaptable processing systems are sought to maintain quality and throughput requirements despite the vulnerabilities of the underlying computational devices. A new approach to autonomous fault-handling which addresses these goals is developed using only a uniplex hardware arrangement. It operates by observing a health metric to achieve Fault Demotion using Recon- figurable Slack (FaDReS). Here an autonomous fault isolation scheme is employed which neither requires test vectors nor suspends the computational throughput, but instead observes the value of a health metric based on runtime input. The deterministic flow of the fault isolation scheme guarantees success in a bounded number of reconfigurations of the FPGA fabric. FaDReS is then extended to the Priority Using Resource Escalation (PURE) online redundancy scheme which considers fault-isolation latency and throughput trade-offs under a dynamic spare arrangement. While deep-submicron designs introduce new challenges, use of adaptive techniques are seen to provide several promising avenues for improving resilience. The scheme developed is demonstrated by hardware design of various signal processing circuits and their implementation on a Xilinx Virtex-4 FPGA device. These include a Discrete Cosine Transform (DCT) core, Motion Estimation (ME) engine, Finite Impulse Response (FIR) Filter, Support Vector Machine (SVM), and Advanced Encryption Standard (AES) blocks in addition to MCNC benchmark circuits. A iii significant reduction in power consumption is achieved ranging from 83% for low motion-activity scenes to 12.5% for high motion activity video scenes in a novel ME engine configuration. For a typical benchmark video sequence, PURE is shown to maintain a PSNR baseline near 32dB. The diagnosability, reconfiguration latency, and resource overhead of each approach is analyzed. Compared to previous alternatives, PURE maintains a PSNR within a difference of 4.02dB to 6.67dB from the fault-free baseline by escalating healthy resources to higher-priority signal processing functions. The results indicate the benefits of priority-aware resiliency over conventional redundancy approaches in terms of fault-recovery, power consumption, and resource-area requirements. Together, these provide a broad range of strategies to achieve autonomous recovery of reconfigurable logic devices under a variety of constraints, operating conditions, and optimization criteria
    corecore