5,134 research outputs found
Recommended from our members
FABRIC: A National-Scale Programmable Experimental Network Infrastructure
FABRIC is a unique national research infrastructure to enable cutting-edge and exploratory research at-scale in networking, cybersecurity, distributed computing and storage systems, machine learning, and science applications. It is an everywhere-programmable nationwide instrument comprised of novel extensible network elements equipped with large amounts of compute and storage, interconnected by high speed, dedicated optical links. It will connect a number of specialized testbeds for cloud research (NSF Cloud testbeds CloudLab and Chameleon), for research beyond 5G technologies (Platforms for Advanced Wireless Research or PAWR), as well as production high-performance computing facilities and science instruments to create a rich fabric for a wide variety of experimental activities
ICONA: Inter Cluster ONOS Network Application
Several Network Operating Systems (NOS) have been proposed in the last few
years for Software Defined Networks; however, a few of them are currently
offering the resiliency, scalability and high availability required for
production environments. Open Networking Operating System (ONOS) is an open
source NOS, designed to be reliable and to scale up to thousands of managed
devices. It supports multiple concurrent instances (a cluster of controllers)
with distributed data stores. A tight requirement of ONOS is that all instances
must be close enough to have negligible communication delays, which means they
are typically installed within a single datacenter or a LAN network. However in
certain wide area network scenarios, this constraint may limit the speed of
responsiveness of the controller toward network events like failures or
congested links, an important requirement from the point of view of a Service
Provider. This paper presents ICONA, a tool developed on top of ONOS and
designed in order to extend ONOS capability in network scenarios where there
are stringent requirements in term of control plane responsiveness. In
particular the paper describes the architecture behind ICONA and provides some
initial evaluation obtained on a preliminary version of the tool.Comment: Paper submitted to a conferenc
When should I use network emulation ?
The design and development of a complex system requires an adequate methodology and efficient instrumental support in order to early detect and correct anomalies in the functional and non-functional properties of the tested protocols. Among the various tools used to provide experimental support for such developments, network emulation relies on real-time production of impairments on real traffic according to a communication model, either realistically or not. This paper aims at simply presenting to newcomers in network emulation (students, engineers, ...) basic principles and practices illustrated with a few commonly used tools. The motivation behind is to fill a gap in terms of introductory and pragmatic papers in this domain. The study particularly considers centralized approaches, allowing cheap and easy implementation in the context of research labs or industrial developments. In addition, an architectural model for emulation systems is proposed, defining three complementary levels, namely hardware, impairment and model levels. With the help of this architectural framework, various existing tools are situated and described. Various approaches for modeling the emulation actions are studied, such as impairment-based scenarios and virtual architectures, real-time discrete simulation and trace-based systems. Those modeling approaches are described and compared in terms of services and we study their ability to respond to various designer needs to assess when emulation is needed
When Should I Use Network Emulation?
The design and development of a complex system requires an adequate
methodology and efficient instrumental support in order to early detect and
correct anomalies in the functional and non-functional properties of the tested
protocols. Among the various tools used to provide experimental support for
such developments, network emulation relies on real-time production of
impairments on real traffic according to a communication model, either
realistically or not.
This paper aims at simply presenting to newcomers in network emulation
(students, engineers, ...) basic principles and practices illustrated with a
few commonly used tools. The motivation behind is to fill a gap in terms of
introductory and pragmatic papers in this domain.
The study particularly considers centralized approaches, allowing cheap and
easy implementation in the context of research labs or industrial developments.
In addition, an architectural model for emulation systems is proposed, defining
three complementary levels, namely hardware, impairment and model levels. With
the help of this architectural framework, various existing tools are situated
and described. Various approaches for modeling the emulation actions are
studied, such as impairment-based scenarios and virtual architectures,
real-time discrete simulation and trace-based systems. Those modeling
approaches are described and compared in terms of services and we study their
ability to respond to various designer needs to assess when emulation is
needed
Fast network configuration in Software Defined Networking
Software Defined Networking (SDN) provides a framework to dynamically adjust and re-program the data plane with the use of flow rules. The realization of highly adaptive SDNs with the ability to respond to changing demands or recover after a network failure in a short period of time, hinges on efficient updates of flow rules. We model the time to deploy a set of flow rules by the update time at the bottleneck switch, and formulate the problem of selecting paths to minimize the deployment time under feasibility constraints as a mixed integer linear program (MILP). To reduce the computation time of determining flow rules, we propose efficient heuristics designed to approximate the minimum-deployment-time solution by relaxing the MILP or selecting the paths sequentially. Through extensive simulations we show that our algorithms outperform current, shortest path based solutions by reducing the total network configuration time up to 55% while having similar packet loss, in the considered scenarios. We also demonstrate that in a networked environment with a certain fraction of failed links, our algorithms are able to reduce the average time to reestablish disrupted flows by 40%
Endpoint-transparent Multipath Transport with Software-defined Networks
Multipath forwarding consists of using multiple paths simultaneously to
transport data over the network. While most such techniques require endpoint
modifications, we investigate how multipath forwarding can be done inside the
network, transparently to endpoint hosts. With such a network-centric approach,
packet reordering becomes a critical issue as it may cause critical performance
degradation.
We present a Software Defined Network architecture which automatically sets
up multipath forwarding, including solutions for reordering and performance
improvement, both at the sending side through multipath scheduling algorithms,
and the receiver side, by resequencing out-of-order packets in a dedicated
in-network buffer.
We implemented a prototype with commonly available technology and evaluated
it in both emulated and real networks. Our results show consistent throughput
improvements, thanks to the use of aggregated path capacity. We give
comparisons to Multipath TCP, where we show our approach can achieve a similar
performance while offering the advantage of endpoint transparency
Deliverable JRA1.1: Evaluation of current network control and management planes for multi-domain network infrastructure
This deliverable includes a compilation and evaluation of available control and management architectures and protocols applicable to a multilayer infrastructure in a multi-domain Virtual Network environment.The scope of this deliverable is mainly focused on the virtualisation of the resources within a network and at processing nodes. The virtualization of the FEDERICA infrastructure allows the provisioning of its available resources to users by means of FEDERICA slices. A slice is seen by the user as a real physical network under his/her domain, however it maps to a logical partition (a virtual instance) of the physical FEDERICA resources. A slice is built to exhibit to the highest degree all the principles applicable to a physical network (isolation, reproducibility, manageability, ...). Currently, there are no standard definitions available for network virtualization or its associated architectures. Therefore, this deliverable proposes the Virtual Network layer architecture and evaluates a set of Management- and Control Planes that can be used for the partitioning and virtualization of the FEDERICA network resources. This evaluation has been performed taking into account an initial set of FEDERICA requirements; a possible extension of the selected tools will be evaluated in future deliverables. The studies described in this deliverable define the virtual architecture of the FEDERICA infrastructure. During this activity, the need has been recognised to establish a new set of basic definitions (taxonomy) for the building blocks that compose the so-called slice, i.e. the virtual network instantiation (which is virtual with regard to the abstracted view made of the building blocks of the FEDERICA infrastructure) and its architectural plane representation. These definitions will be established as a common nomenclature for the FEDERICA project. Other important aspects when defining a new architecture are the user requirements. It is crucial that the resulting architecture fits the demands that users may have. Since this deliverable has been produced at the same time as the contact process with users, made by the project activities related to the Use Case definitions, JRA1 has proposed a set of basic Use Cases to be considered as starting point for its internal studies. When researchers want to experiment with their developments, they need not only network resources on their slices, but also a slice of the processing resources. These processing slice resources are understood as virtual machine instances that users can use to make them behave as software routers or end nodes, on which to download the software protocols or applications they have produced and want to assess in a realistic environment. Hence, this deliverable also studies the APIs of several virtual machine management software products in order to identify which best suits FEDERICA’s needs.Postprint (published version
Evaluación comparativa del rendimiento de controladores SDN de código abierto
Software-defined Networking (SDN) constitutes a new era in the development of internetworking. The SDN paradigm splits the data plane from the control plane. It uses controller equipment, which is responsible for centrally managing several network devices simultaneously. This study analyzes three open-source controllers for SDN based on the OpenFlow protocol. Specifically, the performance of FloodLight, OpenDayLight (ODL), and Ryu controllers is evaluated in terms of latency, throughput, and scalability. In doing so, the Cbench tool is used in an emulated environment with Mininet. The results show that the Ryu controller presents the lowest performance in all the evaluated parameters; ODL provides lower latency and FloodLight higher throughput. Regarding scalability, we conclude that Floodlight can be used in small networks, whereas ODL can be used in dense networks. Furthermore, we evaluate their main characteristics, which must be considered for their choice prior to implementation and deployment.Las redes definidas por software (SDN) constituyen una nueva era en el diseño de la interconexión de redes. El paradigma SDN separa el plano de datos del plano de control. Para esto utiliza un equipo controlador, que se encarga de gestionar de forma centralizada varios dispositivos de red al mismo tiempo. Este estudio analiza tres controladores SDN de código abierto basados en el protocol OpenFlow. Específicamente, el rendimiento de los controladores FloodLight, OpenDayLight (ODL) y Ryu son evaluados en términos de latencia, throughput y escalabilidad. Para ello se utilizó la herramienta Cbench en un entorno emulado con Mininet. Los resultados muestran que el controlador presenta un menor rendimiento en todos los parámetros evaluados; ODL tiene una menor latencia y Floodlight un mayor throughput. En lo que tiene que ver a escalabilidad, se concluye que Floodlight es recomendable para redes pequeñas y ODL para redes densas. Además, evaluamos sus principales características, las cuales deben ser tomadas en cuenta para su elección antes de su implementación y despliegue
- …