7,522 research outputs found

    Resource provisioning in Science Clouds: Requirements and challenges

    Full text link
    Cloud computing has permeated into the information technology industry in the last few years, and it is emerging nowadays in scientific environments. Science user communities are demanding a broad range of computing power to satisfy the needs of high-performance applications, such as local clusters, high-performance computing systems, and computing grids. Different workloads are needed from different computational models, and the cloud is already considered as a promising paradigm. The scheduling and allocation of resources is always a challenging matter in any form of computation and clouds are not an exception. Science applications have unique features that differentiate their workloads, hence, their requirements have to be taken into consideration to be fulfilled when building a Science Cloud. This paper will discuss what are the main scheduling and resource allocation challenges for any Infrastructure as a Service provider supporting scientific applications

    Energy Efficient Virtual Machine Migration in Cloud Data Centers

    Get PDF
    Cloud computing services have been on the rise over the past few decades, which has led to an increase in the number of data centers worldwide which increasingly consume more and more amount of energy for their operation, leading to high carbon dioxide emissions and also high operation costs. Cloud computing infrastructures are designed to support the accessibility and deployment of various service oriented applications by the users. The resources are the major source of the power consumption in data centers along with air conditioning and cooling equipment. Moreover the energy consumption in the cloud is proportional to the resource utilization and data centers are almost the worlds highest consumers of electricity. It is therefore, the need of the hour to devise efficient consolidation schemes for the cloud model to minimize energy and increase Return of Investment(ROI) for the users by decreasing the operating costs. The consolidation problem is NP-complete in nature, which requires heuristic techniques to get a sub-optimal solution. The complexity of the problem increases with increase in cloud infrastructure. We have proposed a new consolidation scheme for the virtual machines(VMs) by improving the host overload detection phase of the scheme. The resulting scheme is effective in reducing the energy and the level of Service Level Agreement(SLA) violations both, to a considerable extent. For testing the performance of our implementation on cloud we need a simulation environment that can provide us an environment with system and behavioural modelling of the actual cloud computing components, and can generate results that can help us in the analysis so that we can deploy them on actual clouds. CloudSim is one such simulation toolkit that allows us to test and analyse our allocation and selection algorithms. In this thesis we have used CloudSim version 3.0.3 to test and analyse our policies and modifications in the current policies. The advantages of using CloudSim 3.0.3 is that it takes very less effort and time to implement cloud-based application and we can test the performance of application services in heterogeneous Cloud environments. The observations are validated by simulating the experiment using the CLoudSim framework and the data provided by PlanetLab

    Computing server power modeling in a data center: survey,taxonomy and performance evaluation

    Full text link
    Data centers are large scale, energy-hungry infrastructure serving the increasing computational demands as the world is becoming more connected in smart cities. The emergence of advanced technologies such as cloud-based services, internet of things (IoT) and big data analytics has augmented the growth of global data centers, leading to high energy consumption. This upsurge in energy consumption of the data centers not only incurs the issue of surging high cost (operational and maintenance) but also has an adverse effect on the environment. Dynamic power management in a data center environment requires the cognizance of the correlation between the system and hardware level performance counters and the power consumption. Power consumption modeling exhibits this correlation and is crucial in designing energy-efficient optimization strategies based on resource utilization. Several works in power modeling are proposed and used in the literature. However, these power models have been evaluated using different benchmarking applications, power measurement techniques and error calculation formula on different machines. In this work, we present a taxonomy and evaluation of 24 software-based power models using a unified environment, benchmarking applications, power measurement technique and error formula, with the aim of achieving an objective comparison. We use different servers architectures to assess the impact of heterogeneity on the models' comparison. The performance analysis of these models is elaborated in the paper

    LEGaTO: first steps towards energy-efficient toolset for heterogeneous computing

    Get PDF
    LEGaTO is a three-year EU H2020 project which started in December 2017. The LEGaTO project will leverage task-based programming models to provide a software ecosystem for Made-in-Europe heterogeneous hardware composed of CPUs, GPUs, FPGAs and dataflow engines. The aim is to attain one order of magnitude energy savings from the edge to the converged cloud/HPC.Peer ReviewedPostprint (author's final draft

    Resource management in a containerized cloud : status and challenges

    Get PDF
    Cloud computing heavily relies on virtualization, as with cloud computing virtual resources are typically leased to the consumer, for example as virtual machines. Efficient management of these virtual resources is of great importance, as it has a direct impact on both the scalability and the operational costs of the cloud environment. Recently, containers are gaining popularity as virtualization technology, due to the minimal overhead compared to traditional virtual machines and the offered portability. Traditional resource management strategies however are typically designed for the allocation and migration of virtual machines, so the question arises how these strategies can be adapted for the management of a containerized cloud. Apart from this, the cloud is also no longer limited to the centrally hosted data center infrastructure. New deployment models have gained maturity, such as fog and mobile edge computing, bringing the cloud closer to the end user. These models could also benefit from container technology, as the newly introduced devices often have limited hardware resources. In this survey, we provide an overview of the current state of the art regarding resource management within the broad sense of cloud computing, complementary to existing surveys in literature. We investigate how research is adapting to the recent evolutions within the cloud, being the adoption of container technology and the introduction of the fog computing conceptual model. Furthermore, we identify several challenges and possible opportunities for future research
    corecore