21,057 research outputs found

    Optimizing Cloud Computing Applications with a Data Center Load Balancing Algorithm

    Get PDF
    Delivering scalable and on-demand computing resources to users through the usage of the cloud has become a common paradigm. The issues of effective resource utilisation and application performance optimisation, however, become more pressing as the demand for cloud services rises. In order to ensure efficient resource allocation and improve application performance, load balancing techniques are essential in dispersing incoming network traffic over several servers. The workload balancing in the context of cloud computing, particularly in the Infrastructure as a Service (IaaS) model, continues to be difficult. Due to available virtual machines and the limited resources, efficient job allocation is essential. To prevent prolonged execution delays or machine breakdowns, cloud service providers must maintain excellent performance and avoid overloading or underloading hosts. The importance of task scheduling in load balancing necessitates compliance with Service Level Agreement (SLA) standards established by cloud developers for consumers. The suggested technique takes into account Quality of Service (QoS) job parameters, VM priorities, and resource allocation in order to maximise resource utilisation and improve load balancing. The proposed load balancing method is in line with the results in the body of existing literature by resolving these problems and the current research gap. According to experimental findings, the Dynamic LBA algorithm currently in use is outperformed by an average resource utilisation of 78%. The suggested algorithm also exhibits excellent performance in terms of accelerated Makespan and decreased execution time

    Autonomic system for optimal resource management in cloud environments

    Full text link
    University of Technology Sydney. Faculty of Engineering and Information Technology.Cloud computing is a large-scale distributed computing paradigm driven by economies of scale, in which a pool of abstracted, virtualized, dynamically-scalable, managed computing power, storage, platforms, and services are delivered on demand to external customers over the Internet. Considering the lack of resources in cloud environments and fluctuating customer demands, cloud providers require to balance their resource load and utilization, and automatically allocate scarce resources to the services in an optimal way to deliver high performance physical and virtual resources and meet Service Level Agreement (SLA) criteria while minimizing their cost. This study proposes an Autonomic System for Optimal Resource Management (AS-ORM) that addresses three main topics of resource management in the cloud environment including: (1) resource estimation, (2) resource discovery and selection, and (3) resource allocation. A fuzzy Workload Prediction (WP) sub-system and a Multi-Objective Task Scheduling optimization (MOTS) sub-system are developed to cover the first two aforementioned topics. The WP sub-systems estimates Virtual Machines’ (VMs’) workload and resource utilization, and predicts Physical Machines’ (PMs) hotspots. The MOTS sub-system determines the optimal pattern to schedule tasks over VMs considering task transfer time, task execution cost/time, the length of the task queue of VMs and power consumption. To optimize the third topic in resource management, resource allocation, VM migration that is the current solution for optimizing physical resources allocation to VMs and load balancing among PMs, is investigated in this study. VM migration has been applied to system load balancing in cloud environments by memory transfer, suspend/resume migration, or live migration for the purpose of minimizing VM downtime and maximizing resource utilization. However, the migration process is both time- and cost-consuming as it requires large size files or memory pages to be transferred, and consumes a huge amount of power and memory for the origin and destination PMs especially for storage VM migration. This process also leads to VM downtime or slowdown. To deal with these shortcomings, a Fuzzy Predictable Task-based System Load Balancing (FP-TBSLB) sub-system is developed that avoids VM migration and achieves system load balancing by transferring extra workload from a poorly performing VM to other compatible VMs with more capacity. To reduce the time factor even more and optimize load balancing over a cloud cluster, FP-TBSLB sub-system applies WP sub-system to not only predict the performance of VMs, but also determine a set of appropriate VMs that have the potential to execute the extra workload imposed on the poorly performing VMs. In addition, FP-TBSLB sub-system employs the MOTS sub-system to migrate the extra workload of poorly performing VMs to the compatible VMs. The AS-ORM system is evaluated using a VMware-vSphere based private cloud environment with VMware ESXi hypervisor. The evaluation results show the benefit of the AS-ORM in reducing the time taken for the load balancing process compared to traditional approaches. The application of this system has the added advantage that the VMs will not be slowed down during the migration process. The system also achieves significant reduction in memory usage, execution time, job makespan and power consumption. Therefore, the AS-ORM dramatically increases VM performance and reduces service response time. The AS-ORM can be applied in the hypervisor layer to optimize resource management and load balancing which boosts the Quality of Service (QoS) expected by cloud customers

    Elastic neural network method for load prediction in cloud computing grid

    Get PDF
    Cloud computing still has no standard definition, yet it is concerned with Internet or network on-demand delivery of resources and services. It has gained much popularity in last few years due to rapid growth in technology and the Internet. Many issues yet to be tackled within cloud computing technical challenges, such as Virtual Machine migration, server association, fault tolerance, scalability, and availability. The most we are concerned with in this research is balancing servers load; the way of spreading the load between various nodes exists in any distributed systems that help to utilize resource and job response time, enhance scalability, and user satisfaction. Load rebalancing algorithm with dynamic resource allocation is presented to adapt with changing needs of a cloud environment. This research presents a modified elastic adaptive neural network (EANN) with modified adaptive smoothing errors, to build an evolving system to predict Virtual Machine load. To evaluate the proposed balancing method, we conducted a series of simulation studies using cloud simulator and made comparisons with previously suggested approaches in the previous work. The experimental results show that suggested method betters present approaches significantly and all these approaches

    Optimizing Virtual Resource Management in Cloud Datacenters

    Get PDF
    Datacenter clouds (e.g., Microsoft\u27s Azure, Google\u27s App Engine, and Amazon\u27s EC2) are emerging as a popular infrastructure for computing and storage due to their high scalability and elasticity. More and more companies and organizations shift their services (e.g., online social networks, Dropbox file hosting) to clouds to avoid large capital expenditures. Cloud systems employ virtualization technology to provide resources in physical machines (PMs) in the form of virtual machines (VMs). Users create VMs deployed on the cloud and each VM consumes resources (e.g., CPU, memory and bandwidth) from its host PM. Cloud providers supply services by signing Service Level Agreement (SLA) with cloud customers that serves as both the blueprint and the warranty for cloud computing. Under-provisioning of resources leads to SLA violations while over-provisioning of resources leads to resource underutilization and then revenue decrease for the cloud providers. Thus, a formidable challenge is effective management of virtual resource to maximize energy efficiency and resource utilization while satisfying the SLA. This proposal is devoted to tackle this challenge by addressing three fundamental and essential issues: i) initial VM allocation, ii) VM migration for load balance, and iii) proactive VM migration for long-term load balance. Accordingly, this proposal consists of three innovative components: (1) Initial Complementary VM Consolidation. Previous resource provisioning strategies either allocate physical resources to virtual machines (VMs) based on static VM resource demands or dynamically handle the variations in VM resource requirements through live VM migrations. However, the former fail to maximize energy efficiency and resource utilization while the latter produce high migration overhead. To handle these problems, we propose an initial VM allocation mechanism that consolidates complementary VMs with spatial/temporal-awareness. Complementary VMs are the VMs whose total demand of each resource dimension (in the spatial space) nearly reaches their host\u27s capacity during VM lifetime period (in the temporal space). Based on our observation of the existence of VM resource utilization patterns, the mechanism predicts the lifetime resource utilization patterns of short-term VMs or periodical resource utilization patterns of long-term VMs. Based on the predicted patterns, it coordinates the requirements of different resources and consolidates complementary VMs in the same physical machine (PM). This mechanism reduces the number of PMs needed to provide VM service hence increases energy efficiency and resource utilization and also reduces the number of VM migrations and SLA violations. (2) Resource Intensity Aware VM Migration for Load Balance. The unique features of clouds pose formidable challenges to achieving effective and efficient load balancing. First, VMs in clouds use different resources (e.g., CPU, bandwidth, memory) to serve a variety of services (e.g., high performance computing, web services, file services), resulting in different overutilized resources in different PMs. Also, the overutilized resources in a PM may vary over time due to the time-varying heterogenous service requests. Second, there is intensive network communication between VMs. However, previous load balancing methods statically assign equal or predefined weights to different resources, which leads to degraded performance in terms of speed and cost to achieve load balance. Also, they do not strive to minimize the VM communications between PMs. This proposed mechanism dynamically assigns different weights to different resources according to their usage intensity in the PM, which significantly reduces the time and cost to achieve load balance and avoids future load imbalance. It also tries to keep frequently communicating VMs in the same PM to reduce bandwidth cost, and migrate VMs to PMs with minimum VM performance degradation. (3) Proactive VM Migration for Long-Term Load Balance. Previous reactive load balancing algorithms migrate VMs upon the occurrence of load imbalance, while previous proactive load balancing algorithms predict PM overload to conduct VM migration. However, both methods cannot maintain long-term load balance and produce high overhead and delay due to migration VM selection and destination PM selection. To overcome these problems, we propose a proactive Markov Decision Process (MDP)-based load balancing algorithm. We handle the challenges of allying MDP in virtual resource management in cloud datacenters, which allows a PM to proactively find an optimal action to transit to a lightly loaded state that will maintain for a longer period of time. We also apply the MDP to determine destination PMs to achieve long-term PM load balance state. Our algorithm reduces the numbers of SLA violations by long-term load balance maintenance, and also reduces the load balancing overhead (e.g., CPU time, energy) and delay by quickly identifying VMs and destination PMs to migrate. Finally, we conducted extensive experiments to evaluate the proposed three mechanisms. i) We conducted simulation experiments based on two real traces and real-world testbed experiments to show that the initial complementary VM consolidation mechanism significantly reduces the number of PMs used, SLA violations and VM migrations of the previous resource provisioning strategies. ii) We conducted trace-driven simulation and real-world testbed experiments to show that RIAL outperforms other load balancing approaches in regards to the number of VM migrations, VM performance degradation and VM communication cost. iii) We conducted trace-driven experiments to show that the MDP-based load balancing algorithm outperforms previous reactive and proactive load balancing algorithms in terms of SLA violation, load balancing efficiency and long-term load balance maintenance

    Load Balancing in Cloud Computing

    Get PDF
    Cloud computing is one of the top trending technologies which primarily focuses on the end user’s use cases. The service provider needs to provide services to many clients. These increasing number of requests from the clients are giving rise to the new inventions in the load scheduling algorithms. There are different scheduling algorithms which are already present in the cloud computing, and some of them includes the Shortest Job First (SJF), First Come First Serve (FCFS), Round Robin (RR) etc. Though there are different parameters to consider when load balancing in cloud computing, makespan (time difference between start time of first task and finish of last task on the same machine) and response time are the most important parameters. This research surveys different load balancing algorithms and aims to improve the SJF load balancing algorithm in cloud computing. In this project, a Modified Shortest Job First (MSJF) and Generalized Priority (GP) load scheduling algorithms are combined to reduce the makespan and optimize the resource utilization. Together, MSJF and GP sends the longest task having high MIPS (million instructions per second) requirements to the machine with a high processing power and the shortest task having low MIPS requirements to the machine with a low processing power. Hence, neither the task with the lowest MIPS requirements nor the task with the highest MIPS requirements needs to wait for a very long time for resource allocation. Every task gets fair priority. Results are shown for SJF, MSJF, and GP in order to compare the different number of tasks using cloud simulator

    Dynamic Virtual Machine Allocation Policy for Load Balancing using Principal Component Analysis and Clustering Technique in Cloud Computing

    Get PDF
    The scalability and agility characteristics of cloud computing allow load balancing to reroute workload requests easily and to enhance overall accessibility. One of the most important services for cloud computing is Infrastructure as a Service (IaaS). There is a large number of physical hosts in a cloud data center for IaaS and it is quite difficult to arrange the allocation of the workload requests manually. Therefore, different load balancing methods have been proposed by researchers to avoid overloaded physical hosts in the cloud data center. However, fewer works have used multivariate analysis in cloud computing environment for considering the dynamic changes of the computing resources. Thus, this work suggests a new Virtual Machine (VM) allocation policy for load balancing by using a multivariate technique, Principal Component Analysis (PCA), and clustering technique. Moreover, PCA and clustering techniques were simulated on a cloud computing simulator, CloudSim. In the proposed allocation policy, a group of VMs were dynamically allocated to physical hosts. The allocation was based on the clusters of hosts according to their similar features in computing resources. The clusters were formed using PCA and a clustering technique based on variables related to the physical hosts such as Million Instructions Per Second (MIPS), Random Access Memory (RAM), bandwidth and storage. The results show that the completion time for all tasks has decreased, and the resource utilization has increased. This will optimize the performance of cloud data centers by effectively utilizing the available resources

    A Review on Various Energy Efficient Techniques in Cloud Environment

    Get PDF
    Cloud computing is web based mostly development and use of engineering. it is a mode of computing within which dynamically scalable and sometimes virtualized resources are provided as a service over the web. Users needn't have data of, experience in, or management over the technology infrastructure "in the cloud" that supports them. programming is one of the core steps to with efficiency exploit the capabilities of heterogeneous computing systems. On cloud computing platform, load equalisation of the whole system will be dynamically handled by using virtualization technology through that it becomes potential to remap virtual machine and physical resources in step with the modification in load. However, so as to boost performance, the virtual machines ought to totally utilize its resources and services by adapting to computing setting dynamically. The load balancing with correct allocation of resources should be bonded so as to boost resource utility and energy efficiency

    ERAM2 - ENERGY BASED RESOURCE ALLOCATION WITH MINIMUM RECKON AND MAXIMUM RECKON

    Get PDF
    The emerging field of cloud computing has flexibility and dominant computational architecture that offers ubiquitous services to users. It is different from traditional architecture because it accommodates resources in a unified way. Due to rapid growth in demands for providing the resources and computation in cloud environments, Resource allocation is considered as primary issues in performance, efficiency, and cost.  For the provisioning of resource, Virtual Machine (VMs) is employed to reduce the response time and executing the tasks according to the available resources.  The users utilize the VMs based on the characteristics of the tasks for effective usage of resources. This helps in load balancing and avoids VMs being in an idle state. Several resource allocation techniques are proposed to maximize the utility of physical resource and minimize the consuming cost of Virtual Machines (VMs). This paper proposes an Energy-Based Resource Allocation with Minimum Reckon and Maximum Reckon (ERAM2); which achieves an efficient scheduling by matching the user tasks on Resource parameters like Accessibility, Availability, Cost, Reliability, Reputation, Response time, Scalability and Throughput in the terms of Maximum Reckon and Minimum Reckon. This paper proposes an Ant Colony - Maximum Reckon and Minimum Reckon (AC-MRMR) method to consolidate all the available resource based on the pheromone value; the score is calculated for each pheromone value. When the score value exceeds Threshold limit then task migration process is carried out for optimized resource allocation of tasks

    Load Balancing and Virtual Machine Allocation in Cloud-based Data Centers

    Get PDF
    As cloud services see an exponential increase in consumers, the demand for faster processing of data and a reliable delivery of services becomes a pressing concern. This puts a lot of pressure on the cloud-based data centers, where the consumers’ data is stored, processed and serviced. The rising demand for high quality services and the constrained environment, make load balancing within the cloud data centers a vital concern. This project aims to achieve load balancing within the data centers by means of implementing a Virtual Machine allocation policy, based on consensus algorithm technique. The cloud-based data center system, consisting of Virtual Machines has been simulated on CloudSim – a Java based cloud simulator
    • …
    corecore