181 research outputs found

    A Logically Centralized Approach for Control and Management of Large Computer Networks

    Get PDF
    Management of large enterprise and Internet Service Provider networks is a complex, error-prone, and costly challenge. It is widely accepted that the key contributors to this complexity are the bundling of control and data forwarding in traditional routers and the use of fully distributed protocols for network control. To address these limitations, the networking research community has been pursuing the vision of simplifying the functional role of a router to its primary task of packet forwarding. This enables centralizing network control at a decision plane where network-wide state can be maintained, and network control can be centrally and consistently enforced. However, scalability and fault-tolerance concerns with physical centralization motivate the need for a more flexible and customizable approach. This dissertation is an attempt at bridging the gap between the extremes of distribution and centralization of network control. We present a logically centralized approach for the design of network decision plane that can be realized by using a set of physically distributed controllers in a network. This approach is aimed at giving network designers the ability to customize the level of control and management centralization according to the scalability, fault-tolerance, and responsiveness requirements of their networks. Our thesis is that logical centralization provides a robust, reliable, and efficient paradigm for management of large networks and we present several contributions to prove this thesis. For network planning, we describe techniques for optimizing the placement of network controllers and provide guidance on the physical design of logically centralized networks. For network operation, algorithms for maintaining dynamic associations between the decision plane and network devices are presented, along with a protocol that allows a set of network controllers to coordinate their decisions, and present a unified interface to the managed network devices. Furthermore, we study the trade-offs in decision plane application design and provide guidance on application state and logic distribution. Finally, we present results of extensive numerical and simulative analysis of the feasibility and performance of our approach. The results show that logical centralization can provide better scalability and fault-tolerance while maintaining performance similarity with traditional distributed approach

    Tree based reliable topology for distributing link state information

    Get PDF
    Finding paths that satisfy the performance requirements of applications according to link state information in a network is known as the Quality-of- Service (QoS) routing problem and has been extensively studied. However, distributing link state information may introduce a significant protocol overhead on network resources. In this thesis, the issue on how to update link state information efficiently and effectively is investigated. A theoretical framework is presented, and a high performance link state policy that is capable of minimizing the false blocking probability of connections under a given update rate constraint is proposed. Through theoretical analysis, it is shown that the proposed policy outperforms the current state of the art in terms of the update rate and higher scalability and reliability

    Foutbestendige toekomstige internetarchitecturen

    Get PDF

    A Survey on the Contributions of Software-Defined Networking to Traffic Engineering

    Get PDF
    Since the appearance of OpenFlow back in 2008, software-defined networking (SDN) has gained momentum. Although there are some discrepancies between the standards developing organizations working with SDN about what SDN is and how it is defined, they all outline traffic engineering (TE) as a key application. One of the most common objectives of TE is the congestion minimization, where techniques such as traffic splitting among multiple paths or advanced reservation systems are used. In such a scenario, this manuscript surveys the role of a comprehensive list of SDN protocols in TE solutions, in order to assess how these protocols can benefit TE. The SDN protocols have been categorized using the SDN architecture proposed by the open networking foundation, which differentiates among data-controller plane interfaces, application-controller plane interfaces, and management interfaces, in order to state how the interface type in which they operate influences TE. In addition, the impact of the SDN protocols on TE has been evaluated by comparing them with the path computation element (PCE)-based architecture. The PCE-based architecture has been selected to measure the impact of SDN on TE because it is the most novel TE architecture until the date, and because it already defines a set of metrics to measure the performance of TE solutions. We conclude that using the three types of interfaces simultaneously will result in more powerful and enhanced TE solutions, since they benefit TE in complementary ways.European Commission through the Horizon 2020 Research and Innovation Programme (GN4) under Grant 691567 Spanish Ministry of Economy and Competitiveness under the Secure Deployment of Services Over SDN and NFV-based Networks Project S&NSEC under Grant TEC2013-47960-C4-3-

    Automated Formal Analysis of Internet Routing Configurations

    Get PDF
    Today\u27s Internet interdomain routing protocol, the Border Gateway Protocol (BGP), is increasingly complicated and fragile due to policy misconfigurations by individual autonomous systems (ASes). To create provably correct networks, the past twenty years have witnessed, among many other efforts, advances in formal network modeling, system verification and testing, and point solutions for network management by formal reasoning. On the conceptual side, the formal models usually abstract away low-level details, specifying what are the correct functionalities but not how to achieve them. On the practical side, system verification of existing networked systems is generally hard, and system testing or simulation provide limited formal guarantees. This is known as a long standing challenge in network practice --- formal reasoning is decoupled from actual implementation. This thesis seeks to bridge formal reasoning and actual network implementation in the setting of the Border Gateway Protocol (BGP), by developing the Formally Verifiable Routing (FVR) toolkit that combines formal methods and programming language techniques. Starting from the formal model, FVR automates verification of routing models and the synthesis of faithful implementations that carries the correctness property. Conversely, starting from large real-world BGP systems with arbitrary policy configurations, automates the analysis of Internet routing configurations, and also includes a novel network reduction technique that scales up existing techniques for automated analysis. By developing the above formal theories and tools, this thesis aims to help network operators to create and manage BGP systems with correctness guarantee

    A scalable heuristic for hybrid IGP/MPLS traffic engineering - Case study on an operational network

    Full text link
    peer reviewedIn current IP networks, a classical way to achieve traffic engineering is to optimise the link metrics. This operation cannot be done too often and can affect the route of a lot of traffic. Multiprotocol Label Switching (MPLS) opens new possibilities to address the limitations of IP systems concerning traffic engineering thanks to explicit label-switched paths (LSPs). This paper proposes a new method based on simulated annealing meta-heuristic to compute a set of LSPs that optimise a given operational objective. The hybrid IGP/MPLS approach takes advantage of both IP and MPLS technologies and provides a flexible method to traffic engineer a network on a day to day basis. We illustrate the capabilities of our method with some simulations and a comparison with other techniques on an existing operational network. The results obtained by setting up a small number of LSPs are nearly optimal and better than by engineering the IGP weights. Moreover, although it could be combined with a static setting of the latter, SAMTE alone gives already the same results as this combination in much less CPU time, which thus allows an administrator to keep its initial and meaningful IGP metrics in his network.DGTRE TOTE

    Fast Local Rerouting for Handling Transient Link Failures

    Full text link

    Optimal route reflection topology design

    Get PDF
    An Autonomous System (AS) is a group of Internet Protocol-based networks with a single and clearly defined external routing policy, usually under single ownership, trust or administrative control. The AS represents a connected group of one or more blocks of IP addresses, called IP prefixes, that have been assigned to that organization and provides a single routing policy to systems outside the AS. The Internet is composed of the interconnection of several thousands of ASes, which use the Border Gateway Protocol (BGP) to exchange network prefixes (aggregations of IP addresses) reachability advertisements. BGP advertisements (or updates) are sent over BGP sessions administratively set between pairs of routers. BGP is a path vector routing protocol and is used to span different ASes. A path vector protocol defines a route as a pairing between a destination and the attributes of the path to that destination. Interior Border Gateway Protocol (iBGP) refers to the BGP neighbor relationship within the same AS. When BGP neighbor relationship are formed between two peers belonging to different AS are called Exterior Border Gateway Protocol (eBGP). In the last case, BGP routers are called Autonomous System Border Routers (ASBRs), while those running only iBGP sessions are referred to as Internal Routers (IRs). Traditional iBGP implementations require a full-mesh of sessions among routers of each AS

    Monitoring Changes in the Stability of Networks Using Eigenvector Centrality

    Get PDF
    Monitoring networks for anomalies is a typical duty of network operators. The conventional monitoring tools available today tend to almost ignore the topological characteristics of the whole network. This thesis takes a different approach from the conventional monitoring tools, by employing the principle of Eigenvector Centrality. Traditionally, this principle is used to analyse vulnerability and social aspects of networks. The proposed model reveals that topological characteristics of a network can be used to improve the conventional unreliability predictors, and to give a better indicator of its potential weaknesses. An effective expected adjacency matrix, k, is introduced in this work to be used with centrality calculations, and it reflects the factors which affect the reliability of a network, for e.g. link downtimes, link metrics, packet loss, etc. Using these calculations, all network backbone routers are assigned values which correspond to the importance of those routers in comparison to the rest of the network nodes. Furthermore, to observe how vulnerable each node could be, nodes are ranked according to the importance values, where the nodes with high ranking values are more vulnerable. This model is able to analyse temporal stability of the network, observing and comparing the rate of change in node ranking values and connectivity caused by the network link failures. The results show that the proposed model is dynamic, and changes according to the dynamics of the topology of the network, i.e. upgrading, link failures, etc.Master i nettverks- og systemadministrasjo
    • …
    corecore