401 research outputs found
Understand Your Chains: Towards Performance Profile-based Network Service Management
Allocating resources to virtualized network functions and services to meet
service level agreements is a challenging task for NFV management and
orchestration systems. This becomes even more challenging when agile
development methodologies, like DevOps, are applied. In such scenarios,
management and orchestration systems are continuously facing new versions of
functions and services which makes it hard to decide how much resources have to
be allocated to them to provide the expected service performance. One solution
for this problem is to support resource allocation decisions with performance
behavior information obtained by profiling techniques applied to such network
functions and services.
In this position paper, we analyze and discuss the components needed to
generate such performance behavior information within the NFV DevOps workflow.
We also outline research questions that identify open issues and missing pieces
for a fully integrated NFV profiling solution. Further, we introduce a novel
profiling mechanism that is able to profile virtualized network functions and
entire network service chains under different resource constraints before they
are deployed on production infrastructure.Comment: Submitted to and accepted by the European Workshop on Software
Defined Networks (EWSDN) 201
Auto-Scaling Network Resources using Machine Learning to Improve QoS and Reduce Cost
Virtualization of network functions (as virtual routers, virtual firewalls,
etc.) enables network owners to efficiently respond to the increasing
dynamicity of network services. Virtual Network Functions (VNFs) are easy to
deploy, update, monitor, and manage. The number of VNF instances, similar to
generic computing resources in cloud, can be easily scaled based on load.
Hence, auto-scaling (of resources without human intervention) has been
receiving attention. Prior studies on auto-scaling use measured network traffic
load to dynamically react to traffic changes. In this study, we propose a
proactive Machine Learning (ML) based approach to perform auto-scaling of VNFs
in response to dynamic traffic changes. Our proposed ML classifier learns from
past VNF scaling decisions and seasonal/spatial behavior of network traffic
load to generate scaling decisions ahead of time. Compared to existing
approaches for ML-based auto-scaling, our study explores how the properties
(e.g., start-up time) of underlying virtualization technology impacts Quality
of Service (QoS) and cost savings. We consider four different virtualization
technologies: Xen and KVM, based on hypervisor virtualization, and Docker and
LXC, based on container virtualization. Our results show promising accuracy of
the ML classifier using real data collected from a private ISP. We report
in-depth analysis of the learning process (learning-curve analysis), feature
ranking (feature selection, Principal Component Analysis (PCA), etc.), impact
of different sets of features, training time, and testing time. Our results
show how the proposed methods improve QoS and reduce operational cost for
network owners. We also demonstrate a practical use-case example
(Software-Defined Wide Area Network (SD-WAN) with VNFs and backbone network) to
show that our ML methods save significant cost for network service leasers
Learning Augmented Optimization for Network Softwarization in 5G
The rapid uptake of mobile devices and applications are posing unprecedented traffic burdens on the existing networking infrastructures. In order to maximize both user experience and investment return, the networking and communications systems are evolving to the next gen- eration – 5G, which is expected to support more flexibility, agility, and intelligence towards provisioned services and infrastructure management. Fulfilling these tasks is challenging, as nowadays networks are increasingly heterogeneous, dynamic and expanded with large sizes. Network softwarization is one of the critical enabling technologies to implement these requirements in 5G. In addition to these problems investigated in preliminary researches about this technology, many new emerging application requirements and advanced opti- mization & learning technologies are introducing more challenges & opportunities for its fully application in practical production environment. This motivates this thesis to develop a new learning augmented optimization technology, which merges both the advanced opti- mization and learning techniques to meet the distinct characteristics of the new application environment. To be more specific, the abstracts of the key contents in this thesis are listed as follows: • We first develop a stochastic solution to augment the optimization of the Network Function Virtualization (NFV) services in dynamical networks. In contrast to the dominant NFV solutions applied for the deterministic networking environments, the inherent network dynamics and uncertainties from 5G infrastructure are impeding the rollout of NFV in many emerging networking applications. Therefore, Chapter 3 investigates the issues of network utility degradation when implementing NFV in dynamical networks, and proposes a robust NFV solution with full respect to the underlying stochastic features. By exploiting the hierarchical decision structures in this problem, a distributed computing framework with two-level decomposition is designed to facilitate a distributed implementation of the proposed model in large-scale networks. • Next, Chapter 4 aims to intertwin the traditional optimization and learning technologies. In order to reap the merits of both optimization and learning technologies but avoid their limitations, promissing integrative approaches are investigated to combine the traditional optimization theories with advanced learning methods. Subsequently, an online optimization process is designed to learn the system dynamics for the network slicing problem, another critical challenge for network softwarization. Specifically, we first present a two-stage slicing optimization model with time-averaged constraints and objective to safeguard the network slicing operations in time-varying networks. Directly solving an off-line solution to this problem is intractable since the future system realizations are unknown before decisions. To address this, we combine the historical learning and Lyapunov stability theories, and develop a learning augmented online optimization approach. This facilitates the system to learn a safe slicing solution from both historical records and real-time observations. We prove that the proposed solution is always feasible and nearly optimal, up to a constant additive factor. Finally, simulation experiments are also provided to demonstrate the considerable improvement of the proposals. • The success of traditional solutions to optimizing the stochastic systems often requires solving a base optimization program repeatedly until convergence. For each iteration, the base program exhibits the same model structure, but only differing in their input data. Such properties of the stochastic optimization systems encourage the work of Chapter 5, in which we apply the latest deep learning technologies to abstract the core structures of an optimization model and then use the learned deep learning model to directly generate the solutions to the equivalent optimization model. In this respect, an encoder-decoder based learning model is developed in Chapter 5 to improve the optimization of network slices. In order to facilitate the solving of the constrained combinatorial optimization program in a deep learning manner, we design a problem-specific decoding process by integrating program constraints and problem context information into the training process. The deep learning model, once trained, can be used to directly generate the solution to any specific problem instance. This avoids the extensive computation in traditional approaches, which re-solve the whole combinatorial optimization problem for every instance from the scratch. With the help of the REINFORCE gradient estimator, the obtained deep learning model in the experiments achieves significantly reduced computation time and optimality loss
z-TORCH: An Automated NFV Orchestration and Monitoring Solution
Autonomous management and orchestration
(MANO) of virtualized resources and services, especially in
large-scale NFV environments, is a big challenge owing to the
stringent delay and performance requirements expected of a
variety of network services. The quality of decision (QoD) of
a MANO system depends on the quality and timeliness of the
information that it receives from the underlying monitoring
system. The data generated by monitoring systems is a
significant contributor to the network and processing load of
MANO systems, impacting thus their performance. This raises
a unique challenge: how to jointly optimize the QoD of MANO
systems while at the same minimizing their monitoring loads at
runtime? This is the main focus of this paper.
In this context we propose a novel automated NFV orchestration
solution called z-TORCH (zero Touch Orchestration)
that jointly optimizes the orchestration and monitoring processes
by exploiting machine learning techniques. The objective is to
enhance the QoD of MANO systems achieving a near-optimal
placement of VNFs at minimum monitoring costs.This work has received funding from
the European Unions Horizon 2020 research and innovation programme under
grant agreement No 761536 (5G-Transformer project
Management And Security Of Multi-Cloud Applications
Single cloud management platform technology has reached maturity and is quite successful in information technology applications. Enterprises and application service providers are increasingly adopting a multi-cloud strategy to reduce the risk of cloud service provider lock-in and cloud blackouts and, at the same time, get the benefits like competitive pricing, the flexibility of resource provisioning and better points of presence. Another class of applications that are getting cloud service providers increasingly interested in is the carriers\u27 virtualized network services. However, virtualized carrier services require high levels of availability and performance and impose stringent requirements on cloud services. They necessitate the use of multi-cloud management and innovative techniques for placement and performance management. We consider two classes of distributed applications – the virtual network services and the next generation of healthcare – that would benefit immensely from deployment over multiple clouds. This thesis deals with the design and development of new processes and algorithms to enable these classes of applications. We have evolved a method for optimization of multi-cloud platforms that will pave the way for obtaining optimized placement for both classes of services. The approach that we have followed for placement itself is predictive cost optimized latency controlled virtual resource placement for both types of applications. To improve the availability of virtual network services, we have made innovative use of the machine and deep learning for developing a framework for fault detection and localization. Finally, to secure patient data flowing through the wide expanse of sensors, cloud hierarchy, virtualized network, and visualization domain, we have evolved hierarchical autoencoder models for data in motion between the IoT domain and the multi-cloud domain and within the multi-cloud hierarchy
- …