6 research outputs found

    On-demand serverless video surveillance with optimal deployment of deep neural networks

    Get PDF
    [EN] We present an approach to optimally deploy Deep Neural Networks (DNNs) in serverless cloud architectures. A serverless architecture allows running code in response to events, automatically managing the required computing resources. However, these resources have limitations in terms of execution environment (CPU only), cold starts, space, scalability, etc. These limitations hinder the deployment of DNNs, especially considering that fees are charged according to the employed resources and the computation time. Our deployment approach is comprised of multiple decoupled software layers that allow effectively managing multiple processes, such as business logic, data access, and computer vision algorithms that leverage DNN optimization techniques. Experimental results in AWS Lambda reveal its potential to build cost-effective ondemand serverless video surveillance systems.This work has been partially supported by the program ELKARTEK 2019 of the Basque Government under project AUTOLIB

    GPU-enabled Function-as-a-Service for Machine Learning Inference

    Full text link
    Function-as-a-Service (FaaS) is emerging as an important cloud computing service model as it can improve the scalability and usability of a wide range of applications, especially Machine-Learning (ML) inference tasks that require scalable resources and complex software configurations. These inference tasks heavily rely on GPUs to achieve high performance; however, support for GPUs is currently lacking in the existing FaaS solutions. The unique event-triggered and short-lived nature of functions poses new challenges to enabling GPUs on FaaS, which must consider the overhead of transferring data (e.g., ML model parameters and inputs/outputs) between GPU and host memory. This paper proposes a novel GPU-enabled FaaS solution that enables ML inference functions to efficiently utilize GPUs to accelerate their computations. First, it extends existing FaaS frameworks such as OpenFaaS to support the scheduling and execution of functions across GPUs in a FaaS cluster. Second, it provides caching of ML models in GPU memory to improve the performance of model inference functions and global management of GPU memories to improve cache utilization. Third, it offers co-designed GPU function scheduling and cache management to optimize the performance of ML inference functions. Specifically, the paper proposes locality-aware scheduling, which maximizes the utilization of both GPU memory for cache hits and GPU cores for parallel processing. A thorough evaluation based on real-world traces and ML models shows that the proposed GPU-enabled FaaS works well for ML inference tasks, and the proposed locality-aware scheduler achieves a speedup of 48x compared to the default, load balancing only schedulers

    Characterizing commodity serverless computing platforms

    Get PDF
    Serverless computing has become a new trending paradigm in cloud computing, allowing developers to focus on the development of core application logic and rapidly construct the prototype via the composition of independent functions. With the development and prosperity of serverless computing, major cloud vendors have successively rolled out their commodity serverless computing platforms. However, the characteristics of these platforms have not been systematically studied. Measuring these characteristics can help developers to select the most adequate serverless computing platform and develop their serverless-based applications in the right way. To fill this knowledge gap, we present a comprehensive study on characterizing mainstream commodity serverless computing platforms, including AWS Lambda, Google Cloud Functions, Azure Functions, and Alibaba Cloud Function Compute. Specifically, we conduct both qualitative analysis and quantitative analysis. In qualitative analysis, we compare these platforms from three aspects (i.e., development, deployment, and runtime) based on their official documentation to construct a taxonomy of characteristics. In quantitative analysis, we analyze the runtime performance of these platforms from multiple dimensions with well-designed benchmarks. First, we analyze three key factors that can influence the startup latency of serverless-based applications. Second, we compare the resource efficiency of different platforms with 16 representative benchmarks. Finally, we measure their performance difference when dealing with different concurrent requests and explore the potential causes in a black-box fashion. Based on the results of both qualitative and quantitative analysis, we derive a series of findings and provide insightful implications for both developers and cloud vendors

    Adapting Microservices in the Cloud with FaaS

    Get PDF
    This project involves benchmarking, microservices and Function-as-a-service (FaaS) across the dimensions of performance and cost. In order to do a comparison this paper proposes a benchmark framework

    Rise of the Planet of Serverless Computing: A Systematic Review

    Get PDF
    Serverless computing is an emerging cloud computing paradigm, being adopted to develop a wide range of software applications. It allows developers to focus on the application logic in the granularity of function, thereby freeing developers from tedious and error-prone infrastructure management. Meanwhile, its unique characteristic poses new challenges to the development and deployment of serverless-based applications. To tackle these challenges, enormous research efforts have been devoted. This paper provides a comprehensive literature review to characterize the current research state of serverless computing. Specifically, this paper covers 164 papers on 17 research directions of serverless computing, including performance optimization, programming framework, application migration, multi-cloud development, testing and debugging, etc. It also derives research trends, focus, and commonly-used platforms for serverless computing, as well as promising research opportunities

    Serverless Computing Strategies on Cloud Platforms

    Full text link
    [ES] Con el desarrollo de la Computaci贸n en la Nube, la entrega de recursos virtualizados a trav茅s de Internet ha crecido enormemente en los 煤ltimos a帽os. Las Funciones como servicio (FaaS), uno de los modelos de servicio m谩s nuevos dentro de la Computaci贸n en la Nube, permite el desarrollo e implementaci贸n de aplicaciones basadas en eventos que cubren servicios administrados en Nubes p煤blicas y locales. Los proveedores p煤blicos de Computaci贸n en la Nube adoptan el modelo FaaS dentro de su cat谩logo para proporcionar computaci贸n basada en eventos altamente escalable para las aplicaciones. Por un lado, los desarrolladores especializados en esta tecnolog铆a se centran en crear marcos de c贸digo abierto serverless para evitar el bloqueo con los proveedores de la Nube p煤blica. A pesar del desarrollo logrado por la inform谩tica serverless, actualmente hay campos relacionados con el procesamiento de datos y la optimizaci贸n del rendimiento en la ejecuci贸n en los que no se ha explorado todo el potencial. En esta tesis doctoral se definen tres estrategias de computaci贸n serverless que permiten evidenciar los beneficios de esta tecnolog铆a para el procesamiento de datos. Las estrategias implementadas permiten el an谩lisis de datos con la integraci贸n de dispositivos de aceleraci贸n para la ejecuci贸n eficiente de aplicaciones cient铆ficas en plataformas cloud p煤blicas y locales. En primer lugar, se desarroll贸 la plataforma CloudTrail-Tracker. CloudTrail-Tracker es una plataforma serverless de c贸digo abierto basada en eventos para el procesamiento de datos que puede escalar autom谩ticamente hacia arriba y hacia abajo, con la capacidad de escalar a cero para minimizar los costos operativos. Seguidamente, se plantea la integraci贸n de GPUs en una plataforma serverless local impulsada por eventos para el procesamiento de datos escalables. La plataforma admite la ejecuci贸n de aplicaciones como funciones severless en respuesta a la carga de un archivo en un sistema de almacenamiento de ficheros, lo que permite la ejecuci贸n en paralelo de las aplicaciones seg煤n los recursos disponibles. Este procesamiento es administrado por un cluster Kubernetes el谩stico que crece y decrece autom谩ticamente seg煤n las necesidades de procesamiento. Ciertos enfoques basados en tecnolog铆as de virtualizaci贸n de GPU como rCUDA y NVIDIA-Docker se eval煤an para acelerar el tiempo de ejecuci贸n de las funciones. Finalmente, se implementa otra soluci贸n basada en el modelo serverless para ejecutar la fase de inferencia de modelos de aprendizaje autom谩tico previamente entrenados, en la plataforma de Amazon Web Services y en una plataforma privada con el framework OSCAR. El sistema crece el谩sticamente de acuerdo con la demanda y presenta una escalado a cero para minimizar los costes. Por otra parte, el front-end proporciona al usuario una experiencia simplificada en la obtenci贸n de la predicci贸n de modelos de aprendizaje autom谩tico. Para demostrar las funcionalidades y ventajas de las soluciones propuestas durante esta tesis se recogen varios casos de estudio que abarcan diferentes campos del conocimiento como la anal铆tica de aprendizaje y la Inteligencia Artificial. Esto demuestra que la gama de aplicaciones donde la computaci贸n serverless puede aportar grandes beneficios es muy amplia. Los resultados obtenidos avalan el uso del modelo serverless en la simplificaci贸n del dise帽o de arquitecturas para el uso intensivo de datos en aplicaciones complejas.[CA] Amb el desenvolupament de la Computaci贸 en el N煤vol, el lliurament de recursos virtualitzats a trav茅s d'Internet ha crescut granment en els 煤ltims anys. Les Funcions com a Servei (FaaS), un dels models de servei m茅s nous dins de la Computaci贸 en el N煤vol, permet el desenvolupament i implementaci贸 d'aplicacions basades en esdeveniments que cobreixen serveis administrats en N煤vols p煤blics i locals. Els prove茂dors de computaci贸 en el N煤vol p煤blic adopten el model FaaS dins del seu cat脿leg per a proporcionar a les aplicacions computaci贸 altament escalable basada en esdeveniments. D'una banda, els desenvolupadors especialitzats en aquesta tecnologia se centren en crear marcs de codi obert serverless per a evitar el bloqueig amb els prove茂dors del N煤vol p煤blic. Malgrat el desenvolupament alcan莽at per la inform脿tica serverless, actualment hi ha camps relacionats amb el processament de dades i l'optimitzaci贸 del rendiment d'execuci贸 en els quals no s'ha explorat tot el potencial. En aquesta tesi doctoral es defineixen tres estrat猫gies inform脿tiques serverless que permeten demostrar els beneficis d'aquesta tecnologia per al processament de dades. Les estrat猫gies implementades permeten l'an脿lisi de dades amb a integraci贸 de dispositius accelerats per a l'execuci贸 eficient d'aplicacion scient铆fiques en plataformes de N煤vol p煤bliques i locals. En primer lloc, es va desenvolupar la plataforma CloudTrail-Tracker. CloudTrail-Tracker 茅s una plataforma de codi obert basada en esdeveniments per al processament de dades serverless que pot escalar autom谩ticament cap amunt i cap avall, amb la capacitat d'escalar a zero per a minimitzar els costos operatius. A continuaci贸 es planteja la integraci贸 de GPUs en una plataforma serverless local impulsada per esdeveniments per al processament de dades escalables. La plataforma admet l'execuci贸 d'aplicacions com funcions severless en resposta a la c脿rrega d'un arxiu en un sistema d'emmagatzemaments de fitxers, la qual cosa permet l'execuci贸 en paral路lel de les aplicacions segon sels recursos disponibles. Este processament 茅s administrat per un cluster Kubernetes el脿stic que creix i decreix autom脿ticament segons les necessitats de processament. Certs enfocaments basats en tecnologies de virtualitzaci贸 de GPU com rCUDA i NVIDIA-Docker s'avaluen per a accelerar el temps d'execuci贸 de les funcions. Finalment s'implementa una altra soluci贸 basada en el model serverless per a executar la fase d'infer猫ncia de models d'aprenentatge autom脿tic pr猫viament entrenats en la plataforma de Amazon Web Services i en una plataforma privada amb el framework OSCAR. El sistema creix el脿sticament d'acord amb la demanda i presenta una escalada a zero per a minimitzar els costos. D'altra banda el front-end proporciona a l'usuari una experi猫ncia simplificada en l'obtenci贸 de la predicci贸 de models d'aprenentatge autom脿tic. Per a demostrar les funcionalitats i avantatges de les solucions proposades durant esta tesi s'arrepleguen diversos casos d'estudi que comprenen diferents camps del coneixement com l'anal铆tica d'aprenentatge i la Intel路lig猫ncia Artificial. Aix貌 demostra que la gamma d'aplicacions on la computaci贸 serverless pot aportar grans beneficis 茅s molt 脿mplia. Els resultats obtinguts avalen l'煤s del model serverless en la simplificaci贸 del disseny d'arquitectures per a l'煤s intensiu de dades en aplicacions complexes.[EN] With the development of Cloud Computing, the delivery of virtualized resources over the Internet has greatly grown in recent years. Functions as a Service (FaaS), one of the newest service models within Cloud Computing, allows the development and implementation of event-based applications that cover managed services in public and on-premises Clouds. Public Cloud Computing providers adopt the FaaS model within their catalog to provide event-driven highly-scalable computing for applications. On the one hand, developers specialized in this technology focus on creating open-source serverless frameworks to avoid the lock-in with public Cloud providers. Despite the development achieved by serverless computing, there are currently fields related to data processing and execution performance optimization where the full potential has not been explored. In this doctoral thesis three serverless computing strategies are defined that allow to demonstrate the benefits of this technology for data processing. The implemented strategies allow the analysis of data with the integration of accelerated devices for the efficient execution of scientific applications on public and on-premises Cloud platforms. Firstly, the CloudTrail-Tracker platform was developed to extract and process learning analytics in the Cloud. CloudTrail-Tracker is an event-driven open-source platform for serverless data processing that can automatically scale up and down, featuring the ability to scale to zero for minimizing the operational costs. Next, the integration of GPUs in an event-driven on-premises serverless platform for scalable data processing is discussed. The platform supports the execution of applications as serverless functions in response to the loading of a file in a file storage system, which allows the parallel execution of applications according to available resources. This processing is managed by an elastic Kubernetes cluster that automatically grows and shrinks according to the processing needs. Certain approaches based on GPU virtualization technologies such as rCUDA and NVIDIA-Docker are evaluated to speed up the execution time of the functions. Finally, another solution based on the serverless model is implemented to run the inference phase of previously trained machine learning models on theAmazon Web Services platform and in a private platform with the OSCAR framework. The system grows elastically according to demand and is scaled to zero to minimize costs. On the other hand, the front-end provides the user with a simplified experience in obtaining the prediction of machine learning models. To demonstrate the functionalities and advantages of the solutions proposed during this thesis, several case studies are collected covering different fields of knowledge such as learning analytics and Artificial Intelligence. This shows the wide range of applications where serverless computing can bring great benefits. The results obtained endorse the use of the serverless model in simplifying the design of architectures for the intensive data processing in complex applications.Naranjo Delgado, DM. (2021). Serverless Computing Strategies on Cloud Platforms [Tesis doctoral]. Universitat Polit猫cnica de Val猫ncia. https://doi.org/10.4995/Thesis/10251/160916TESI
    corecore