4 research outputs found

    Dynamic Network Slicing Using Deep Reinforcement Learning

    Get PDF
    Nowadays network slicing is one of the biggest drivers of new elements in the 5G network business. This is because this paradigm allows the creation of independent slices, with their virtually and logically separated radio, network and computational resources. Using network slicing, operators sell infrastructure resources of any kind to tenants, while tenants use these resources to sell services to their customers, the end users. In this context, a problem that is essential to solve is how to improve the operator’s profit, ensuring compliance with the requests’ SLAs and distributing network resources in order to increase its usage rate. This dissertation proposes to design two algorithms based on DRL for slice admission in the transport network, learning which request to accept and reject while guaranteeing the requirements of the tenants requests. The contributions of this study start with the formalization of the problem of slice admission, followed by its simulation and implementation of DRL agents using Containernet, the Ryu controller, OpenAI Gym and the PyTorch framework. The result is two DRL-based algorithms capable of achieving good performances in this simulated scenario.Atualmente o network slicing é um dos maiores potenciadores de novos elementos no negócio das redes 5G. Isto deve-se ao facto de este paradigma permitir a criação de slices independentes, com os seus recursos rádio, de rede e computacionais virtual e logicamente separados. Utilizando network slicing, as operadoras poderão vender recursos de infraestrutura de qualquer tipo a tenants. Os tenants utilizam estes recursos para vender serviços aos seus clientes, os utilizadores finais. Neste contexto, um problema que é fundamental resolver é o de como melhorar o lucro da operadora, garantindo o cumprimento dos SLAs dos pedidos e distribuindo os recursos da rede de forma a aumentar a sua utilização. Nesta dissertação propõe-se desenhar dois algoritmos baseados em DRL para a admissão de slices na rede de transporte, aprendendo que pedidos aceitar e rejeitar, procurando satisfazer sempre os requisitos dos pedidos dos tenants. Os contributos deste estudo passam pela formalização do problema da admissão de slices na rede, seguindo-se a sua simulação e implementação dos agentes utilizando conjuntamente o Containernet, o controlador Ryu, o OpenAI Gym e o framework PyTorch. O resultado são dois algoritmos baseados em DRL capazes de atingir boas performances neste cenário simulado

    Performance modelling for scalable deep learning

    Get PDF
    Performance modelling for scalable deep learning is very important to quantify the efficiency of large parallel workloads. Performance models are used to obtain run-time estimates by modelling various aspects of an application on a target system. Designing performance models requires comprehensive analysis in order to build accurate models. Limitations of current performance models include poor explainability in the computation time of the internal processes of a neural network model and limited applicability to particular architectures. Existing performance models in deep learning have been proposed, which are broadly categorized into two methodologies: analytical modelling and empirical modelling. Analytical modelling utilizes a transparent approach that involves converting the internal mechanisms of the model or applications into a mathematical model that corresponds to the goals of the system. Empirical modelling predicts outcomes based on observation and experimentation, characterizes algorithm performance using sample data, and is a good alternative to analytical modelling. However, both these approaches have limitations, such as poor explainability in the computation time of the internal processes of a neural network model and poor generalisation. To address these issues, hybridization of the analytical and empirical approaches has been applied, leading to the development of a novel generic performance model that provides a general expression of a deep neural network framework in a distributed environment, allowing for accurate performance analysis and prediction. The contributions can be summarized as follows: In the initial study, a comprehensive literature review led to the development of a performance model based on synchronous stochastic gradient descent (S-SGD) for analysing the execution time performance of deep learning frameworks in a multi-GPU environment. This model’s evaluation involved three deep learning models (Convolutional Neural Networks (CNN), Autoencoder (AE), and Multilayer Perceptron (MLP)), implemented in three popular deep learning frameworks (MXNet, Chainer, and TensorFlow) respectively, with a focus on following an analytical approach. Additionally, a generic expression for the performance model was formulated, considering intrinsic parameters and extrinsic scaling factors that impact computing time in a distributed environment. This formulation involved a global optimization problem with a cost function dependent on unknown constants within the generic expression. Differential evolution was utilized to identify the best fitting values, matching experimentally determined computation times. Furthermore, to enhance the accuracy and stability of the performance model, regularization techniques were applied. Lastly, the proposed generic performance model underwent experimental evaluation in a real-world application. The results of this evaluation provided valuable insights into the influence of hyperparameters on performance, demonstrating the robustness and applicability of the performance model in understanding and optimizing model behavior

    Performance Analysis of Distributed and Scalable Deep Learning

    No full text
    With renewed global interest for Artificial Intelligence (AI) methods, the past decade has seen a myriad of new programming models and tools that enable better and faster Machine Learning (ML). More recently, a subset of ML known as Deep Learning (DL) raised an increased interest due to its inherent ability to tackle efficiently novel cognitive computing applications. DL allows computational models that are composed of multiple processing layers to learn in an automated way representations of data with multiple levels of abstraction, and can deliver higher predictive accuracy when trained on larger data sets. Based on Artificial Neural Networks (ANN), DL is now at the core of state of the art voice recognition systems (which enable easy control over e.g. Internet-of- Things (IoT) smart home appliances for instance), self-driving car engine, online recommendation systems. The ecosystem of DL frameworks is fast evolving, as well as the DL architectures that are shown to perform well on specialized tasks and to exploit GPU accelerators. For this reason, the frequent performance evaluation of the DL ecosystem is re- quired, especially since the advent of novel distributed training frameworks such as Horovod allowing for scalable training across multiple computing resources. In this paper, the scalability evaluation of the reference DL frameworks (Tensorflow, Keras, MXNet, and PyTorch) is performed over up-to-date High Performance Comput- ing (HPC) resources to compare the efficiency of differ- ent implementations across several hardware architectures (CPU and GPU). Experimental results demonstrate that the DistributedDataParallel features in the Pytorch library seem to be the most efficient framework for distributing the training process across many devices, allowing to reach a throughput speedup of 10.11 when using 12 NVidia Tesla V100 GPUs when training Resnet44 on the CIFAR10 dataset
    corecore