124 research outputs found
E2: a framework for NFV applications
By moving network appliance functionality from proprietary
hardware to software, Network Function Virtualization
promises to bring the advantages of cloud computing to
network packet processing. However, the evolution of cloud
computing (particularly for data analytics) has greatly bene-
fited from application-independent methods for scaling and
placement that achieve high efficiency while relieving programmers
of these burdens. NFV has no such general management
solutions. In this paper, we present a scalable and
application-agnostic scheduling framework for packet processing,
and compare its performance to current approaches
Self-managing cloud-native applications : design, implementation and experience
Running applications in the cloud efficiently requires much more than deploying software in virtual machines. Cloud applications have to be continuously managed: (1) to adjust their resources to the incoming load and (2) to face transient failures replicating and restarting components to provide resiliency on unreliable infrastructure. Continuous management monitors application and infrastructural metrics to provide automated and responsive reactions to failures (health management) and changing environmental conditions (auto-scaling) minimizing human intervention.
In the current practice, management functionalities are provided as infrastructural or third party services. In both cases they are external to the application deployment. We claim that this approach has intrinsic limits, namely that separating management functionalities from the application prevents them from naturally scaling with the application and requires additional management code and human intervention. Moreover, using infrastructure provider services for management functionalities results in vendor lock-in effectively preventing cloud applications to adapt and run on the most effective cloud for the job.
In this paper we discuss the main characteristics of cloud native applications, propose a novel architecture that enables scalable and resilient self-managing applications in the cloud, and relate on our experience in porting a legacy application to the cloud applying cloud-native principles
A Survey on Data Plane Programming with P4: Fundamentals, Advances, and Applied Research
With traditional networking, users can configure control plane protocols to
match the specific network configuration, but without the ability to
fundamentally change the underlying algorithms. With SDN, the users may provide
their own control plane, that can control network devices through their data
plane APIs. Programmable data planes allow users to define their own data plane
algorithms for network devices including appropriate data plane APIs which may
be leveraged by user-defined SDN control. Thus, programmable data planes and
SDN offer great flexibility for network customization, be it for specialized,
commercial appliances, e.g., in 5G or data center networks, or for rapid
prototyping in industrial and academic research. Programming
protocol-independent packet processors (P4) has emerged as the currently most
widespread abstraction, programming language, and concept for data plane
programming. It is developed and standardized by an open community and it is
supported by various software and hardware platforms. In this paper, we survey
the literature from 2015 to 2020 on data plane programming with P4. Our survey
covers 497 references of which 367 are scientific publications. We organize our
work into two parts. In the first part, we give an overview of data plane
programming models, the programming language, architectures, compilers,
targets, and data plane APIs. We also consider research efforts to advance P4
technology. In the second part, we analyze a large body of literature
considering P4-based applied research. We categorize 241 research papers into
different application domains, summarize their contributions, and extract
prototypes, target platforms, and source code availability.Comment: Submitted to IEEE Communications Surveys and Tutorials (COMS) on
2021-01-2
Elastic Highly Available Cloud Computing
High availability and elasticity are two the cloud computing services technical features. Elasticity is a key feature of cloud computing where provisioning of resources is closely tied to the runtime demand. High availability assure that cloud applications are resilient to failures. Existing cloud solutions focus on providing both features at the level of the virtual resource through virtual machines by managing their restart, addition, and removal as needed. These existing solutions map applications to a specific design, which is not suitable for many applications especially virtualized telecommunication applications that are required to meet carrier grade standards. Carrier grade applications typically rely on the underlying platform to manage their availability by monitoring heartbeats, executing recoveries, and attempting repairs to bring the system back to normal. Migrating such applications to the cloud can be particularly challenging, especially if the elasticity policies target the application only, without considering the underlying platform contributing to its high availability (HA). In this thesis, a Network Function Virtualization (NFV) framework is introduced; the challenges and requirements of its use in mobile networks are discussed. In particular, an architecture for NFV framework entities in the virtual environment is proposed. In order to reduce signaling traffic congestion and achieve better performance, a criterion to bundle multiple functions of virtualized evolved packet-core in a single physical device or a group of adjacent devices is proposed. The analysis shows that the proposed grouping can reduce the network control traffic by 70 percent. Moreover, a comprehensive framework for the elasticity of highly available applications that considers the elastic deployment of the platform and the HA placement of the application’s components is proposed. The approach is applied to an internet protocol multimedia subsystem (IMS) application and demonstrate how, within a matter of seconds, the IMS application can be scaled up while maintaining its HA status
Software Defined Applications in Cellular and Optical Networks
abstract: Small wireless cells have the potential to overcome bottlenecks in wireless access through the sharing of spectrum resources. A novel access backhaul network architecture based on a Smart Gateway (Sm-GW) between the small cell base stations, e.g., LTE eNBs, and the conventional backhaul gateways, e.g., LTE Servicing/Packet Gateways (S/P-GWs) has been introduced to address the bottleneck. The Sm-GW flexibly schedules uplink transmissions for the eNBs. Based on software defined networking (SDN) a management mechanism that allows multiple operator to flexibly inter-operate via multiple Sm-GWs with a multitude of small cells has been proposed. This dissertation also comprehensively survey the studies that examine the SDN paradigm in optical networks. Along with the PHY functional split improvements, the performance of Distributed Converged Cable Access Platform (DCCAP) in the cable architectures especially for the Remote-PHY and Remote-MACPHY nodes has been evaluated. In the PHY functional split, in addition to the re-use of infrastructure with a common FFT module for multiple technologies, a novel cross functional split interaction to cache the repetitive QAM symbols across time at the remote node to reduce the transmission rate requirement of the fronthaul link has been proposed.Dissertation/ThesisDoctoral Dissertation Electrical Engineering 201
Design and Implementation of a Distributed Mobility Management Entity (MME) on OpenStack
Network Functions Virtualisation (NFV) involves the implementation of network functions, for example firewalls and routers, as software applications that can run on general-purpose servers. In present-day networks, each network function is typically implemented on dedicated and proprietary hardware. By utilising virtualisation technologies, NFV enables network functions to be deployed on cloud computing infrastructure in data centers.
This thesis discusses the application of NFV to the Evolved Packet Core (EPC) in Long Term Evolution (LTE) networks; specifically to the Mobility Management Entity (MME), a control plane entity in the EPC. With the convergence of cloud computing and mobile networks, conventional architectures of network elements need to be re-designed in order to fully harness benefits such as scalability and elasticity. To this end, we design and implement a distributed MME with a three-tier architecture common to web applications. We highlight design considerations for moving MME functionality to the cloud and compare our new distributed design to that of a standalone MME. We deploy and test the distributed MME on two separate OpenStack clouds. Our results indicate that the benefits of scalability and resilience can outweigh the marginal increase in latency for EPC procedures. We find that the latency is dependent on the actual placement of MME components within the data center. Also, we believe that extensions to the OpenStack platform are required before it can meet performance and availability requirements for telecommunication applications
Modeling and Dimensioning of a Virtualized MME for 5G Mobile Networks
Network function virtualization is considered one of
the key technologies for developing future mobile networks. In this
paper, we propose a theoretical framework to evaluate the performance of a Long-Term Evolution (LTE) virtualized mobility management entity (vMME) hosted in a data center. This theoretical
framework consists of 1) a queuing network to model the vMME
in a data center and 2) analytic expressions to estimate the overall
mean system delay and the signaling workload to be processed by
the vMME. We validate our mathematical model by simulation.
One direct use of the proposed model is vMME dimensioning, i.e.,
to compute the number of vMME processing instances to provide
a target system delay given the number of users in the system.
Additionally, the paper includes a scalability analysis of the system. In our study, we consider the billing model and a data center
setup of Amazon Elastic Compute Cloud service and estimate the
processing time of MME processing instances for different LTE
control procedures experimentally. For the considered setup, our
results show that the vMME is scalable for signaling workloads
up to 37 000 LTE control procedures per second for a target mean
system delay of 1 ms. The system design and database performance
assumed imposes this limit in the system scalability.This work was supported in part by the Spanish Ministry of Economy
and Competitiveness and the European Regional Development Fund (project
TIN2013-46223-P) and in part by the Spanish Ministry of Education, Culture,
and Sport under FPU Grant 13/04833
- …