12 research outputs found

    Rule-based SLA management for revenue maximisation in cloud computing markets

    Get PDF
    This paper introduces several Business Rules for maximising the revenue of Providers in Cloud Computing Markets. These rules apply in both negotiation and execution time, and enforce the achievement of Business-Level Objectives by establishing a bidirectional data flow between market and resource layers. The experiments demonstrate that the revenue is maximized by using both resource data when negotiating, and economic information when managing the resources.Postprint (published version

    Mitigating the Effects of Partial Resource Failures for Cloud Providers

    Get PDF
    Competition for users on a global market is fierce, forcing enterprises to provide for better, faster services while offering the same more cheaply. At the same time, users choose to remain oblivious of the infrastructure behind the service – only demanding that it works. Cloud service failures and inefficient management of such failures can result in significant financial cost, loss of reputation for providers, and drive key customers away. At the same time failure situations can never be completely avoided. To mitigate their effects we present a decision model for providers to help them decide which jobs to keep running and which to cancel in order to minimize loss of revenue and key customers during partial resource failures. The results of the evaluation of the model and its extension show its ability to significantly improve revenue. Furthermore the model can also help to reduce the number of cancelled jobs

    CLOUD SERVICE REVENUE MANAGEMENT

    Get PDF
    Successful Internet service offerings can only thrive if customers are satisfied with service performance. While large service providers can usually cope with fluctuations of customer visits retaining acceptable Quality of Service, small and medium-sizes enterprises face a big challenge due to limited resources in the IT infrastructure. Popular services, such as justin.tv and SmugMug, rely on external resources provided by cloud computing providers in order to satisfy their customers demands at all times. The paradigm of cloud computing refers to the delivery model of computing services as a utility in a pay-as-you-go manner. In this paper, we provide and computationally evaluate decision models and policies that can help cloud computing providers increase their revenue under the realistic assumption of scarce resources and under both informational certainty and uncertainty of customers? resource requirement predictions. Our results show that in both cases under certainty and under uncertainty applying the dynamic pricing policy significantly increases revenue while using the client classification policy substantially reduces revenue. We also show that, for all policies, the presence of uncertainty causes losses in revenue; when the client classification policy is applied, losses can even amount to more than 8%

    Maximising revenue in cloud computing markets by means of economically enhanced SLA management

    Get PDF
    This paper proposes a bidirectional communication between market brokers and resource managers in Cloud Computing Markets. This communication is implemented by means of an Economically Enhanced Resource Manager (EERM), that supports the negotiation process by deciding which tasks can be allocated or not, and under which economic and technical conditions. The EERM also uses the economic information that collects from market layers to manage the resources accordingly to concrete BLOs. This paper shows several Business Policies and Rules for maximizing the revenue of a Cloud Provider that sells its services and resources in a market. Their validity is demonstrated through several experiments that shown how the application of these rules can have a positive influence in the revenue and minimize the violations of Service-Level Agreements.Preprin

    Management of Cloud Infastructures: Policy-Based Revenue Optimization

    Get PDF
    Competition on global markets forces many enterprises to make use of new applications, reduce process times and at the same time cut the costs of their IT-infrastructure. To achieve this, it is necessary to maintain a high degree of flexibility with respect to the IT-infrastructure. Facing this challenge the idea of Cloud computing has been gaining interest lately. Cloud services can be accessed on demand without knowledge of the underlying infrastructure and have already succeeded in helping companies deploy products faster. Using Cloud services the New York Times managed to convert scanned images containing 11 million articles into PDF within 24 hours at a cost of merely 240 US-$. However Cloud providers will only offer their services, if they can realize sufficient benefit. To achieve this, the efficiency of Cloud infrastructure management must be increased. To this end we propose the use of concepts from revenue management and further enhancements

    AUGURES : profit-aware web infrastructure management

    Get PDF
    Over the last decade, advances in technology together with the increasing use of the Internet for everyday tasks, are causing profound changes in end-users, as well as in businesses and technology providers. The widespread adoption of high-speed and ubiquitous Internet access, is also changing the way users interact with Web applications and their expectations in terms of Quality-of-Service (QoS) and User eXperience (UX). Recently, Cloud computing has been rapidly adopted to host and manage Web applications, due to its inherent cost effectiveness and on-demand scaling of infrastructures. However, system administrators still need to make manual decisions about the parameters that affect the business results of their applications ie., setting QoS targets and defining metrics for scaling the number of servers during the day. Therefore, understanding the workload and user behavior ¿the demand, poses new challenges for capacity planning and scalability ¿the supply, and ultimately for the success of a Web site. This thesis contributes to the current state-of-art of Web infrastructure management by providing: i) a methodology for predicting Web session revenue; ii) a methodology to determine high response time effect on sales; and iii) a policy for profit-aware resource management, that relates server capacity, to QoS, and sales. The approach leverages Machine Learning (ML) techniques on custom, real-life datasets from an Ecommerce retailer featuring popular Web applications. Where the experimentation shows how user behavior and server performance models can be built from offline information, to determine how demand and supply relations work as resources are consumed. Producing in this way, economical metrics that are consumed by profit-aware policies, that allow the self-configuration of cloud infrastructures to an optimal number of servers under a variety of conditions. While at the same time, the thesis, provides several insights applicable for improving Autonomic infrastructure management and the profitability of Ecommerce applications.Durante la última década, avances en tecnología junto al incremento de uso de Internet, están causando cambios en los usuarios finales, así como también a las empresas y proveedores de tecnología. La adopción masiva del acceso ubicuo a Internet de alta velocidad, crea cambios en la forma de interacción con las aplicaciones Web y en las expectativas de los usuarios en relación de calidad de servicio (QoS) y experiencia de usuario (UX) ofrecidas. Recientemente, el modelo de computación Cloud ha sido adoptado rápidamente para albergar y gestionar aplicaciones Web, debido a su inherente efectividad en costos y servidores bajo demanda. Sin embargo, los administradores de sistema aún tienen que tomar decisiones manuales con respecto a los parámetros de ejecución que afectan a los resultados de negocio p.ej. definir objetivos de QoS y métricas para escalar en número de servidores. Por estos motivos, entender la carga y el comportamiento de usuario (la demanda), pone nuevos desafíos a la planificación de capacidad y escalabilidad (el suministro), y finalmente el éxito de un sitio Web.Esta tesis contribuye al estado del arte actual en gestión de infraestructuras Web presentado: i) una metodología para predecir los beneficios de una sesión Web; ii) una metodología para determinar el efecto de tiempos de respuesta altos en las ventas; y iii) una política para la gestión de recursos basada en beneficios, al relacionar la capacidad de los servidores, QoS, y ventas. La propuesta se basa en aplicar técnicas Machine Learning (ML) a fuentes de datos de producción de un proveedor de Ecommerce, que ofrece aplicaciones Web populares. Donde los experimentos realizados muestran cómo modelos de comportamiento de usuario y de rendimiento de servidor pueden obtenerse de datos históricos; con el fin de determinar la relación entre la demanda y el suministro, según se utilizan los recursos. Produciendo así, métricas económicas que son luego aplicadas en políticas basadas en beneficios, para permitir la auto-configuración de infraestructuras Cloud a un número adecuado de servidores. Mientras que al mismo tiempo, la tesis provee información relevante para mejorar la gestión de infraestructuras Web de forma autónoma y aumentar los beneficios en aplicaciones de Ecommerce

    Improved self-management of datacenter systems applying machine learning

    Get PDF
    Autonomic Computing is a Computer Science and Technologies research area, originated during mid 2000's. It focuses on optimization and improvement of complex distributed computing systems through self-control and self-management. As distributed computing systems grow in complexity, like multi-datacenter systems in cloud computing, the system operators and architects need more help to understand, design and optimize manually these systems, even more when these systems are distributed along the world and belong to different entities and authorities. Self-management lets these distributed computing systems improve their resource and energy management, a very important issue when resources have a cost, by obtaining, running or maintaining them. Here we propose to improve Autonomic Computing techniques for resource management by applying modeling and prediction methods from Machine Learning and Artificial Intelligence. Machine Learning methods can find accurate models from system behaviors and often intelligible explanations to them, also predict and infer system states and values. These models obtained from automatic learning have the advantage of being easily updated to workload or configuration changes by re-taking examples and re-training the predictors. So employing automatic modeling and predictive abilities, we can find new methods for making "intelligent" decisions and discovering new information and knowledge from systems. This thesis departs from the state of the art, where management is based on administrators expertise, well known data, ad-hoc studied algorithms and models, and elements to be studied from computing machine point of view; to a novel state of the art where management is driven by models learned from the same system, providing useful feedback, making up for incomplete, missing or uncertain data, from a global network of datacenters point of view. - First of all, we cover the scenario where the decision maker works knowing all pieces of information from the system: how much will each job consume, how is and will be the desired quality of service, what are the deadlines for the workload, etc. All of this focusing on each component and policy of each element involved in executing these jobs. -Then we focus on the scenario where instead of fixed oracles that provide us information from an expert formula or set of conditions, machine learning is used to create these oracles. Here we look at components and specific details while some part of the information is not known and must be learned and predicted. - We reduce the problem of optimizing resource allocations and requirements for virtualized web-services to a mathematical problem, indicating each factor, variable and element involved, also all the constraints the scheduling process must attend to. The scheduling problem can be modeled as a Mixed Integer Linear Program. Here we face an scenario of a full datacenter, further we introduce some information prediction. - We complement the model by expanding the predicted elements, studying the main resources (this is CPU, Memory and IO) that can suffer from noise, inaccuracy or unavailability. Once learning predictors for certain components let the decision making improve, the system can become more ¿expert-knowledge independent¿ and research can focus on an scenario where all the elements provide noisy, uncertainty or private information. Also we introduce to the management optimization new factors as for each datacenter context and costs may change, turning the model as "multi-datacenter" - Finally, we review of the cost of placing datacenters depending on green energy sources, and distribute the load according to green energy availability

    An adaptive admission control and load balancing algorithm for a QoS-aware Web system

    Get PDF
    The main objective of this thesis focuses on the design of an adaptive algorithm for admission control and content-aware load balancing for Web traffic. In order to set the context of this work, several reviews are included to introduce the reader in the background concepts of Web load balancing, admission control and the Internet traffic characteristics that may affect the good performance of a Web site. The admission control and load balancing algorithm described in this thesis manages the distribution of traffic to a Web cluster based on QoS requirements. The goal of the proposed scheduling algorithm is to avoid situations in which the system provides a lower performance than desired due to servers' congestion. This is achieved through the implementation of forecasting calculations. Obviously, the increase of the computational cost of the algorithm results in some overhead. This is the reason for designing an adaptive time slot scheduling that sets the execution times of the algorithm depending on the burstiness that is arriving to the system. Therefore, the predictive scheduling algorithm proposed includes an adaptive overhead control. Once defined the scheduling of the algorithm, we design the admission control module based on throughput predictions. The results obtained by several throughput predictors are compared and one of them is selected to be included in our algorithm. The utilisation level that the Web servers will have in the near future is also forecasted and reserved for each service depending on the Service Level Agreement (SLA). Our load balancing strategy is based on a classical policy. Hence, a comparison of several classical load balancing policies is also included in order to know which of them better fits our algorithm. A simulation model has been designed to obtain the results presented in this thesis

    Advanced Concepts for Renewable Energy Supply of Data Centres

    Get PDF
    The rapid increase of cloud computing, high performance computing (HPC) and the vast growth in Internet and Social Media use have aroused the interest in energy consumption and the carbon footprint of Data Centres. Data Centres primarily contain electronic equipment used for data processing (servers), data storage (storage equipment), and communications (network equipment). Collectively, this equipment processes, stores, and transmits digital information and is known as information technology (IT) equipment. Advanced Concepts for Renewable Energy Supply of Data Centres introduces a number of technical solutions for the supply of power and cooling energy into Data Centres with enhanced utilisation of renewable energy sources in order to achieve low energy Data Centres. Because of the high energy density nature of these unique infrastructures, it is essential to implement energy efficiency measures and reduce consumption before introducing any renewable energy source. A holistic approach is used with the objective of integrating many technical solutions such as management of the IT (Information Technology) load, efficient electrical supply to the IT systems, Low-Ex air-conditioning systems, interaction with district heating and cooling networks, re-use of heat, free cooling (air, seawater, groundwater), optimal use of heat and cold storage, electrical storage and integration in smart grids. This book is therefore a catalogue of advanced technical concepts that could be integrated into Data Centres portfolio in order to increase the overall efficiency and the share of renewable energies in power and cooling supply. Based on dynamic energy models implemented in TRNSYS some concepts are deeply evaluated through yearly simulations. The results of the simulation are illustrated with Sankey charts, where the energy flows per year within the subsystems of each concept for a selected scenario are shown, and graphs showing the results of parametric analysis. A set of environmental metrics (as the non-renewable primary energy) and financial metrics (CAPEX and OPEX) as well of energy efficiency metrics like the well-known PUE, are described and used to evaluate the different technical concepts

    Advanced Concepts for Renewable Energy Supply of Data Centres

    Get PDF
    The rapid increase of cloud computing, high performance computing (HPC) and the vast growth in Internet and Social Media use have aroused the interest in energy consumption and the carbon footprint of Data Centres. Data Centres primarily contain electronic equipment used for data processing (servers), data storage (storage equipment), and communications (network equipment). Collectively, this equipment processes, stores, and transmits digital information and is known as information technology (IT) equipment. Advanced Concepts for Renewable Energy Supply of Data Centres introduces a number of technical solutions for the supply of power and cooling energy into Data Centres with enhanced utilisation of renewable energy sources in order to achieve low energy Data Centres. Because of the high energy density nature of these unique infrastructures, it is essential to implement energy efficiency measures and reduce consumption before introducing any renewable energy source. A holistic approach is used with the objective of integrating many technical solutions such as management of the IT (Information Technology) load, efficient electrical supply to the IT systems, Low-Ex air-conditioning systems, interaction with district heating and cooling networks, re-use of heat, free cooling (air, seawater, groundwater), optimal use of heat and cold storage, electrical storage and integration in smart grids. This book is therefore a catalogue of advanced technical concepts that could be integrated into Data Centres portfolio in order to increase the overall efficiency and the share of renewable energies in power and cooling supply. Based on dynamic energy models implemented in TRNSYS some concepts are deeply evaluated through yearly simulations. The results of the simulation are illustrated with Sankey charts, where the energy flows per year within the subsystems of each concept for a selected scenario are shown, and graphs showing the results of parametric analysis. A set of environmental metrics (as the non-renewable primary energy) and financial metrics (CAPEX and OPEX) as well of energy efficiency metrics like the well-known PUE, are described and used to evaluate the different technical concepts
    corecore