390 research outputs found

    Scheduling Stochastic Multi-Stage Jobs to Elastic Hybrid Cloud Resources

    Full text link
    [EN] We consider a special workflow scheduling problem in a hybrid-cloud-based workflow management system in which tasks are linearly dependent, compute-intensive, stochastic, deadline-constrained and executed on elastic and distributed cloud resources. This kind of problems closely resemble many real-time and workflow-based applications. Three optimization objectives are explored: number, usage time and utilization of rented VMs. An iterated heuristic framework is presented to schedule jobs event by event which mainly consists of job collecting and event scheduling. Two job collecting strategies are proposed and two timetabling methods are developed. The proposed methods are calibrated through detailed designs of experiments and sound statistical techniques. With the calibrated components and parameters, the proposed algorithm is compared to existing methods for related problems. Experimental results show that the proposal is robust and effective for the problems under study.This work is sponsored by the National Natural Science Foundations of China (Nos. 71401079, 61572127, 61472192), the National Key Research and Development Program of China (No. 2017YFB1400801) and the Collaborative Innovation Center of Wireless Communications Technology. Ruben Ruiz is partially supported by the Spanish Ministry of Economy and Competitiveness, under the project "SCHEYARD-Optimization of Scheduling Problems in Container Yards" (No. DPI2015-65895-R) financed by FEDER funds.Zhu, J.; Li, X.; Ruiz García, R.; Xu, X. (2018). Scheduling Stochastic Multi-Stage Jobs to Elastic Hybrid Cloud Resources. IEEE Transactions on Parallel and Distributed Systems. 29(6):1401-1415. https://doi.org/10.1109/TPDS.2018.2793254S1401141529

    Energy and Performance: Management of Virtual Machines: Provisioning, Placement, and Consolidation

    Get PDF
    Cloud computing is a new computing paradigm that offers scalable storage and compute resources to users on demand through Internet. Public cloud providers operate large-scale data centers around the world to handle a large number of users request. However, data centers consume an immense amount of electrical energy that can lead to high operating costs and carbon emissions. One of the most common and effective method in order to reduce energy consumption is Dynamic Virtual Machines Consolidation (DVMC) enabled by the virtualization technology. DVMC dynamically consolidates Virtual Machines (VMs) into the minimum number of active servers and then switches the idle servers into a power-saving mode to save energy. However, maintaining the desired level of Quality-of-Service (QoS) between data centers and their users is critical for satisfying users’ expectations concerning performance. Therefore, the main challenge is to minimize the data center energy consumption while maintaining the required QoS. This thesis address this challenge by presenting novel DVMC approaches to reduce the energy consumption of data centers and improve resource utilization under workload independent quality of service constraints. These approaches can be divided into three main categories: heuristic, meta-heuristic and machine learning. Our first contribution is a heuristic algorithm for solving the DVMC problem. The algorithm uses a linear regression-based prediction model to detect over-loaded servers based on the historical utilization data. Then it migrates some VMs from the over-loaded servers to avoid further performance degradations. Moreover, our algorithm consolidates VMs on fewer number of server for energy saving. The second and third contributions are two novel DVMC algorithms based on the Reinforcement Learning (RL) approach. RL is interesting for highly adaptive and autonomous management in dynamic environments. For this reason, we use RL to solve two main sub-problems in VM consolidation. The first sub-problem is the server power mode detection (sleep or active). The second sub-problem is to find an effective solution for server status detection (overloaded or non-overloaded). The fourth contribution of this thesis is an online optimization meta-heuristic algorithm called Ant Colony System-based Placement Optimization (ACS-PO). ACS is a suitable approach for VM consolidation due to the ease of parallelization, that it is close to the optimal solution, and its polynomial worst-case time complexity. The simulation results show that ACS-PO provides substantial improvement over other heuristic algorithms in reducing energy consumption, the number of VM migrations, and performance degradations. Our fifth contribution is a Hierarchical VM management (HiVM) architecture based on a three-tier data center topology which is very common use in data centers. HiVM has the ability to scale across many thousands of servers with energy efficiency. Our sixth contribution is a Utilization Prediction-aware Best Fit Decreasing (UP-BFD) algorithm. UP-BFD can avoid SLA violations and needless migrations by taking into consideration the current and predicted future resource requirements for allocation, consolidation, and placement of VMs. Finally, the seventh and the last contribution is a novel Self-Adaptive Resource Management System (SARMS) in data centers. To achieve scalability, SARMS uses a hierarchical architecture that is partially inspired from HiVM. Moreover, SARMS provides self-adaptive ability for resource management by dynamically adjusting the utilization thresholds for each server in data centers.Siirretty Doriast

    CDOXplorer: Simulation-based genetic optimization of software deployment and reconfiguration in the cloud

    Get PDF
    Migrating existing enterprise software to cloud platforms involves the comparison of various cloud deployment options (CDOs). A CDO comprises a combination of a specific cloud environment, deployment architecture, and runtime reconfiguration rules for dynamic resource scaling. Our simulator CDOSim can evaluate CDOs, e.g., regarding response times and costs. However, the design space to be searched for well-suited solutions is very large. In this paper, we approach this optimization problem with the novel genetic algorithm CDOXplorer. It uses techniques of the search-based software engineering field and simulations with CDOSim to assess the fitness of CDOs. An experimental evaluation that employs, among others, the cloud environments Amazon EC2 and Microsoft Windows Azure, shows that CDOXplorer can find solutions that surpass those of other state-of-the-art techniques by up to 60\%. Our experiment code and data and an implementation of CDOXplorer are available as open source software

    QoS-aware Cloud Infrastructure Provisioning in Heterogeneous Environments

    Get PDF
    Over the last decades Information Technology (IT) has become an enabler for nearly all businesses from industrial production to finance. The IT resources required for these business activities are usually provided by local and remote data centers. Although most resources are still hosted in companies’ proprietary data centers, cloud computing initiated a paradigm shift in IT service provisioning from owning to leasing resources and services. Today, over 50% of German companies use cloud services while shifting services into the cloud has become an emerging trend. Cloud computing, which is often referred to as the fifth utility in addition to water, electricity, gas, and telephony, provides commoditized computation resources that are available any time on demand in the required quantity. However, in contrast to other commodities, a single quality level is insufficient for IT service provisioning. Instead, the required quality for a provided IT service depends on the various functional and non-functional requirements. For example, highly interactive applications such as cloud gaming require a high quality level in terms of latency. Providers of cloud services have to face a highly competitive market. Cost advantages in cloud computing are primarily achieved by utilizing large centralized data centers at low-cost locations. However, this kind of resource provisioning impacts the quality of service of different types of services such as the aforementioned interactive multimedia services that possess strict quality of service constraints. Hence, infrastructure providers have to face a trade-off between cost reduction and adherence to the required Quality of Service (QoS) attributes. Apart from how services are provisioned, the way of consuming IT services also changed substantially over the last years. Mobile devices have begun to replace locally installed desktop computers at an accelerated pace. By utilizing these mobile devices, service providers are confronted with two major challenges: (i) a cellular network connection, which potentially causes a higher and more fluctuating latency and (ii) severely limited resources compared to local Personal Computers (PCs). These two aspects restrict the utilization of multimedia services, e. g., cloud gaming. To address these challenges, we present two novel approaches for (i) resource planning on a global level for multiple services with heterogeneous QoS characteristics and (ii) the augmentation of the centralized cloud infrastructure with locally installed resources to provide viable multimedia services to mobile devices. As the first major contribution, we introduce the Cloud Data Center Selection Problem (CDCSP). This problem describes the data center placement and resource selection on a global scale. We consider the role of a cloud provider, who aims to dimension resources in a cost-minimal fashion under the consideration of multiple services with different QoS attributes. Based on a mathematical optimization model, we propose the exact solution approach CDCSP-EXA.KOM. Due to the high complexity and the resulting computational effort to find the optimal solution, we propose and analyze four heuristic approaches to identify the most appropriate one for the given problem. As a first heuristic, we propose an approach based on linear program relaxation, CDCSP-REL.KOM. Furthermore, to take the specific structure of the problem into consideration, we develop the custom tailored CDCSP-PBST.KOM approach, which is based on a prioritized processing of demands and supplies. To further improve the results, we combine multiple heuristics to a Best-of-Breed approach, named CDCSP-BoB. KOM. Finally, as a metaheuristic improvement procedure, we propose the tabu search approach CDCSP-TS.KOM. To assess the practical applicability and performance of these optimization approaches, we analyze them in detail and compare their performance in a quantitatively. The second major contribution of this work addresses the augmentation of the centralized cloud infrastructure with local resources to provide services to mobile devices. Therefore, we formulate the Dynamic Cloudlet Placement and Selection Problem (DCPSP), as a multi-period resource planning problem, which includes local characteristics, such as space for hosting resources and available network bandwidth. We focus on a cloud provider who aims to augment the centralized infrastructure using local resources to improve the QoS guarantees for mobile used applications. We formalize the problem as a mathematical optimization model and derive the exact solution approach DCPSP-EXA.KOM. Due to the high complexity that is caused by an optimization over many time slots, we propose the heuristic optimization approach DCPSP-HEU.KOM. We assess the performance of these two approaches by the means of quantitative evaluation. In summary, the contributions of this thesis provide the means for a cost-efficient and QoS-aware resource selection in cloud infrastructures. We contribute the formalization of the problems and algorithms to support the efficient planning of future cloud infrastructures in environments with a multitude of heterogeneous services on a global scale. Furthermore, to enable mobile users to consume multimedia cloud services, we propose an optimization model and algorithms to augment a global centralized infrastructure by local resource units

    How to Place Your Apps in the Fog -- State of the Art and Open Challenges

    Full text link
    Fog computing aims at extending the Cloud towards the IoT so to achieve improved QoS and to empower latency-sensitive and bandwidth-hungry applications. The Fog calls for novel models and algorithms to distribute multi-service applications in such a way that data processing occurs wherever it is best-placed, based on both functional and non-functional requirements. This survey reviews the existing methodologies to solve the application placement problem in the Fog, while pursuing three main objectives. First, it offers a comprehensive overview on the currently employed algorithms, on the availability of open-source prototypes, and on the size of test use cases. Second, it classifies the literature based on the application and Fog infrastructure characteristics that are captured by available models, with a focus on the considered constraints and the optimised metrics. Finally, it identifies some open challenges in application placement in the Fog

    Smart Decision-Making via Edge Intelligence for Smart Cities

    Get PDF
    Smart cities are an ambitious vision for future urban environments. The ultimate aim of smart cities is to use modern technology to optimize city resources and operations while improving overall quality-of-life of its citizens. Realizing this ambitious vision will require embracing advancements in information communication technology, data analysis, and other technologies. Because smart cities naturally produce vast amounts of data, recent artificial intelligence (AI) techniques are of interest due to their ability to transform raw data into insightful knowledge to inform decisions (e.g., using live road traffic data to control traffic lights based on current traffic conditions). However, training and providing these AI applications is non-trivial and will require sufficient computing resources. Traditionally, cloud computing infrastructure have been used to process computationally intensive tasks; however, due to the time-sensitivity of many of these smart city applications, novel computing hardware/technologies are required. The recent advent of edge computing provides a promising computing infrastructure to support the needs of the smart cities of tomorrow. Edge computing pushes compute resources close to end users to provide reduced latency and improved scalability — making it a viable candidate to support smart cities. However, it comes with hardware limitations that are necessary to consider. This thesis explores the use of the edge computing paradigm for smart city applications and how to make efficient, smart decisions related to their available resources. This is done while considering the quality-of-service provided to end users. This work can be seen as four parts. First, this work touches on how to optimally place and serve AI-based applications on edge computing infrastructure to maximize quality-of-service to end users. This is cast as an optimization problem and solved with efficient algorithms that approximate the optimal solution. Second, this work investigates the applicability of compression techniques to reduce offloading costs for AI-based applications in edge computing systems. Finally, this thesis then demonstrate how edge computing can support AI-based solutions for smart city applications, namely, smart energy and smart traffic. These applications are approached using the recent paradigm of federated learning. The contributions of this thesis include the design of novel algorithms and system design strategies for placement and scheduling of AI-based services on edge computing systems, formal formulation for trade-offs between delivered AI model performance and latency, compression for offloading decisions for communication reductions, and evaluation of federated learning-based approaches for smart city applications

    A general deep reinforcement learning hyperheuristic framework for solving combinatorial optimization problems

    Get PDF
    Many problem-specific heuristic frameworks have been developed to solve combinatorial optimization problems, but these frameworks do not generalize well to other problem domains. Metaheuristic frameworks aim to be more generalizable compared to traditional heuristics, however their performances suffer from poor selection of low-level heuristics (operators) during the search process. An example of heuristic selection in a metaheuristic framework is the adaptive layer of the popular framework of Adaptive Large Neighborhood Search (ALNS). Here, we propose a selection hyperheuristic framework that uses Deep Reinforcement Learning (Deep RL) as an alternative to the adaptive layer of ALNS. Unlike the adaptive layer which only considers heuristics’ past performance for future selection, a Deep RL agent is able to take into account additional information from the search process, e.g., the difference in objective value between iterations, to make better decisions. This is due to the representation power of Deep Learning methods and the decision making capability of the Deep RL agent which can learn to adapt to different problems and instance characteristics. In this paper, by integrating the Deep RL agent into the ALNS framework, we introduce Deep Reinforcement Learning Hyperheuristic (DRLH), a general framework for solving a wide variety of combinatorial optimization problems and show that our framework is better at selecting low-level heuristics at each step of the search process compared to ALNS and a Uniform Random Selection (URS). Our experiments also show that while ALNS can not properly handle a large pool of heuristics, DRLH is not negatively affected by increasing the number of heuristics.publishedVersio

    Energy Aware Resource Allocation for Clouds Using Two Level Ant Colony Optimization

    Get PDF
    In cloud environment resources are dynamically allocated, adjusted, and deallocated. When to allocate and how many resources to allocate is a challenging task. Resources allocated optimally and at the right time not only improve the utilization of resources but also increase energy efficiency, provider's profit and customers' satisfaction. This paper presents ant colony optimization (ACO) based energy aware solution for resource allocation problem. The proposed energy aware resource allocation (EARA) methodology strives to optimize allocation of resources in order to improve energy efficiency of the cloud infrastructure while satisfying quality of service (QoS) requirements of the end users. Resources are allocated to jobs according to their QoS requirements. For energy efficient and QoS aware allocation of resources, EARA uses ACO at two levels. First level ACO allocates Virtual Machines (VMs) resources to jobs whereas second level ACO allocates Physical Machines (PMs) resources to VMs. Server consolidation and dynamic performance scaling of PMs are employed to conserve energy. The proposed methodology is implemented in CloudSim and the results are compared with existing popular resource allocation methods. Simulation results demonstrate that EARA achieves desired QoS and superior energy gains through better utilization of resources. EARA outperforms major existing resource allocation methods and achieves up to 10.56 % saving in energy consumption
    corecore