14,543 research outputs found

    Performance Improvement of Cloud Computing Data Centers Using Energy Efficient Task Scheduling Algorithms

    Full text link
    Cloud computing is a technology that provides a platform for the sharing of resources such as software, infrastructure, application and other information. It brings a revolution in Information Technology industry by offering on-demand of resources. Clouds are basically virtualized datacenters and applications offered as services. Data center hosts hundreds or thousands of servers which comprised of software and hardware to respond the client request. A large amount of energy requires to perform the operation.. Cloud Computing is facing lot of challenges like Security of Data, Consumption of energy, Server Consolidation, etc. The research work focuses on the study of task scheduling management in a cloud environment. The main goal is to improve the performance (resource utilization and redeem the consumption of energy) in data centers. Energy-efficient scheduling of workloads helps to redeem the consumption of energy in data centers, thus helps in better USAge of resource. This is further reducing operational costs and provides benefits to the clients and also to cloud service provider. In this abstract of paper, the task scheduling in data centers have been compared. Cloudsim a toolkit for modeling and simulation of cloud computing environment has been used to implement and demonstrate the experimental results. The results aimed at analyzing the energy consumed in data centers and shows that by having reduce the consumption of energy the cloud productivity can be improved

    Energy-Aware Lease Scheduling in Virtualized Data Centers

    Full text link
    Energy efficiency has become an important measurement of scheduling algorithms in virtualized data centers. One of the challenges of energy-efficient scheduling algorithms, however, is the trade-off between minimizing energy consumption and satisfying quality of service (e.g. performance, resource availability on time for reservation requests). We consider resource needs in the context of virtualized data centers of a private cloud system, which provides resource leases in terms of virtual machines (VMs) for user applications. In this paper, we propose heuristics for scheduling VMs that address the above challenge. On performance evaluation, simulated results have shown a significant reduction on total energy consumption of our proposed algorithms compared with an existing First-Come-First-Serve (FCFS) scheduling algorithm with the same fulfillment of performance requirements. We also discuss the improvement of energy saving when additionally using migration policies to the above mentioned algorithms.Comment: 10 pages, 2 figures, Proceedings of the Fifth International Conference on High Performance Scientific Computing, March 5-9, 2012, Hanoi, Vietna

    Energy-Efficient Management of Data Center Resources for Cloud Computing: A Vision, Architectural Elements, and Open Challenges

    Full text link
    Cloud computing is offering utility-oriented IT services to users worldwide. Based on a pay-as-you-go model, it enables hosting of pervasive applications from consumer, scientific, and business domains. However, data centers hosting Cloud applications consume huge amounts of energy, contributing to high operational costs and carbon footprints to the environment. Therefore, we need Green Cloud computing solutions that can not only save energy for the environment but also reduce operational costs. This paper presents vision, challenges, and architectural elements for energy-efficient management of Cloud computing environments. We focus on the development of dynamic resource provisioning and allocation algorithms that consider the synergy between various data center infrastructures (i.e., the hardware, power units, cooling and software), and holistically work to boost data center energy efficiency and performance. In particular, this paper proposes (a) architectural principles for energy-efficient management of Clouds; (b) energy-efficient resource allocation policies and scheduling algorithms considering quality-of-service expectations, and devices power usage characteristics; and (c) a novel software technology for energy-efficient management of Clouds. We have validated our approach by conducting a set of rigorous performance evaluation study using the CloudSim toolkit. The results demonstrate that Cloud computing model has immense potential as it offers significant performance gains as regards to response time and cost saving under dynamic workload scenarios.Comment: 12 pages, 5 figures,Proceedings of the 2010 International Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA 2010), Las Vegas, USA, July 12-15, 201

    Secure data sharing in cloud computing: a comprehensive review

    Get PDF
    Cloud Computing is an emerging technology, which relies on sharing computing resources. Sharing of data in the group is not secure as the cloud provider cannot be trusted. The fundamental difficulties in distributed computing of cloud suppliers is Data Security, Sharing, Resource scheduling and Energy consumption. Key-Aggregate cryptosystem used to secure private/public data in the cloud. This key is consistent size aggregate for adaptable decisions of ciphertext in cloud storage. Virtual Machines (VMs) provisioning is effectively empowered the cloud suppliers to effectively use their accessible resources and get higher benefits. The most effective method to share information resources among the individuals from the group in distributed storage is secure, flexible and efficient. Any data stored in different cloud data centers are corrupted, recovery using regenerative coding. Security is provided many techniques like Forward security, backward security, Key-Aggregate cryptosystem, Encryption and Re-encryption etc. The energy is reduced using Energy-Efficient Virtual Machines Scheduling in Multi-Tenant Data Centers

    SCORE: Simulator for cloud optimization of resources andenergy consumption

    Get PDF
    Achieving efficiency both in terms of resource utilisation and energy consumption is acomplex challenge, especially in large-scale wide-purpose data centers that serve cloud- computing services. Simulation presents an appropriate solution for the development andtesting of strategies that aim to improve efficiency problems before their applications inproduction environments. Various cloud simulators have been proposed to cover differentaspects of the operation environment of cloud-computing systems. In this paper, we define the SCORE tool, which is dedicated to the simulation of energy-efficient monolithicand parallel-scheduling models and for the execution of heterogeneous, realistic and synthetic workloads. The simulator has been evaluated through empirical tests. The results ofthe experiments confirm that SCORE is a performant and reliable tool for testing energy- efficiency, security, and scheduling strategies in cloud-computing environments.European Cooperation in Science and Technology (COST) COST Action IC140

    Energy-Efficient Load Balancing Algorithm for Workflow Scheduling in Cloud Data Centers Using Queuing and Thresholds

    Get PDF
    Cloud computing is a rapidly growing technology that has been implemented in various fields in recent years, such as business, research, industry, and computing. Cloud computing provides different services over the internet, thus eliminating the need for personalized hardware and other resources. Cloud computing environments face some challenges in terms of resource utilization, energy efficiency, heterogeneous resources, etc. Tasks scheduling and virtual machines (VMs) are used as consolidation techniques in order to tackle these issues. Tasks scheduling has been extensively studied in the literature. The problem has been studied with different parameters and objectives. In this article, we address the problem of energy consumption and efficient resource utilization in virtualized cloud data centers. The proposed algorithm is based on task classification and thresholds for efficient scheduling and better resource utilization. In the first phase, workflow tasks are pre-processed to avoid bottlenecks by placing tasks with more dependencies and long execution times in separate queues. In the next step, tasks are classified based on the intensities of the required resources. Finally, Particle Swarm Optimization (PSO) is used to select the best schedules. Experiments were performed to validate the proposed technique. Comparative results obtained on benchmark datasets are presented. The results show the effectiveness of the proposed algorithm over that of the other algorithms to which it was compared in terms of energy consumption, makespan, and load balancing

    Designing Parametric Constraint Based Power Aware Scheduling System in a Virtualized Cloud Environment

    Get PDF
    The increasing rate of the demand for computational resources has led to the production of largescale data centers. They consume huge amounts of electrical power resulting in high operational costs and carbon dioxide emissions. Power-related costs have become one of the major economic factors in IT data-centers, and companies and the research community are currently working on new efficient power aware resource management strategies, also known as 201C;Green IT201D;. Here we propose a framework for autonomic scheduling of tasks based upon some parametric constraints. In this paper we propose an analysis of the critical factors affecting the energy consumption of cloud servers in cloud computing and consideration to make performance very fast by using Sigar API to solve speed problems. In PCBPAS we impose some parametric constraints during task allocation to the server that can be adjusted dynamically to balance the server2019;s workloads in an efficient way so that CPU consumption can be improved and energy saving be achieved

    Data Center Load Forecast Using Dependent Mixture Model

    Get PDF
    The dependency on cloud computing is increasing day by day. With the boom of data centers, the cost is also increasing, which forces industries to come up with techniques and methodologies to reduce the data center energy use. Load forecasting plays a vital role in both efficient scheduling and operating a data center as a virtual power plant. In this thesis work a stochastic method, based on dependent mixtures is developed to model the data center load and is used for day-ahead forecast. The method is validated using three data sets from National Renewable Energy Laboratory (NREL) and one other data centers. The proposed method proved better than the classical autoregressive, moving-average, as well as the neural network-based forecasting method, and resulted in a reduction of 7.91% mean absolute percentage error (MAPE) for the forecast. A more accurate forecast can improve power scheduling and resource management reducing the variable cost of power generation as well as the overall data center operating cost, which was quantified as a yearly savings of $13,705 for a typical 100 MW coal fired tier-IV data center

    Improved Cloud resource allocation: how INDIGO-Datacloud is overcoming the current limitations in Cloud schedulers

    Get PDF
    Trabajo presentado a: 22nd International Conference on Computing in High Energy and Nuclear Physics (CHEP2016) 10–14 October 2016, San Francisco.Performing efficient resource provisioning is a fundamental aspect for any resource provider. Local Resource Management Systems (LRMS) have been used in data centers for decades in order to obtain the best usage of the resources, providing their fair usage and partitioning for the users. In contrast, current cloud schedulers are normally based on the immediate allocation of resources on a first-come, first-served basis, meaning that a request will fail if there are no resources (e.g. OpenStack) or it will be trivially queued ordered by entry time (e.g. OpenNebula). Moreover, these scheduling strategies are based on a static partitioning of the resources, meaning that existing quotas cannot be exceeded, even if there are idle resources allocated to other projects. This is a consequence of the fact that cloud instances are not associated with a maximum execution time and leads to a situation where the resources are under-utilized. These facts have been identified by the INDIGO-DataCloud project as being too simplistic for accommodating scientific workloads in an efficient way, leading to an underutilization of the resources, a non desirable situation in scientific data centers. In this work, we will present the work done in the scheduling area during the first year of the INDIGO project and the foreseen evolutions.The authors want to acknowledge the support of the INDIGO-DataCloud (grant number 653549) project, funded by the European Commission’s Horizon 2020 Framework Programme.Peer Reviewe
    corecore