11 research outputs found

    A Modified Black Hole-Based Task Scheduling Technique for Cloud Computing Environment

    Get PDF
    The issue of scheduling is one of the most important ones to be considered by providers of the cloud computing in the data center. Using a suitable solution lets the providers of cloud computing use the available resources more. Additionally, the satisfaction of clients is met through provision of service quality parameters. Most of the solutions for this problem aim at one of the service quality factors and in order to achieve this goal, variety of methods are used. Using the algorithm of modified black hole in this paper, a proper solution is presented to tackle the problem of scheduling the affairs in cloud environment. The proposed method reduces makespan, increases degree of load balancing, and improves the resource`s utilization by considering the capability of each virtual machine. We have compared the proposed algorithm with existing task scheduling algorithms. Simulation results indicate that the proposed algorithm makes a good improvement regarding the makespan and amount of resource utilization compared to schedulers based on Random assignment and particle swarm optimization Algorithms

    Scheduling Algorithms for Cloud: A Survey and Analysis

    Get PDF
    Cloud Computing is a fast growing computing paradigm due to the vast benefits it provides to the users. Scheduling becomes one of the key aspects due to the pay-as-you-go nature of the Cloud. The factors affecting the technique of scheduling applied change with change in scenarios. For instance for scheduling in hybrid clouds the data transfer speed has to be taken into consideration whereas for mobile environments scheduling becomes dependent on context change. Moreover scheduling can be improvised on many fronts such as energy efficiency, cost minimization, Maximization of resource utilization, etc. This paper surveys scheduling techniques in various Cloud Computing scenarios and sites the most efficient scheduling technique available for a particular set of user needs by comparing various techniques and the problems they address

    Improved Task Scheduling for Virtual Machines in the Cloud based on the Gravitational Search Algorithm

    Full text link
    The rapid and convenient provision of the available computing resources is a crucial requirement in modern cloud computing environments. However, if only the execution time is taken into account when the resources are scheduled, it could lead to imbalanced workloads as well as to significant under-utilisation of the involved Virtual Machines (VMs). In the present work a novel task scheduling scheme is introduced, which is based on the proper adaptation of a modern and quite effective evolutionary optimization method, the Gravitational Search Algorithm (GSA). The proposed scheme aims at optimizing the entire scheduling procedure, in terms of both the tasks execution time and the system (VMs) resource utilisation. Moreover, the fitness function was properly selected considering both the above factors in an appropriately weighted function in order to obtain better results for large inputs. Sufficient simulation experiments show the efficiency of the proposed scheme, as well as its excellence over related approaches of the bibliography, with similar objectives.Comment: 8 page

    BIG, MEDIUM AND LITTLE (BML) SCHEDULING IN FOG ENVIRONMENT

    Get PDF
    BIG, MEDIUM AND LITTLE (BML) SCHEDULING IN FOG ENVIRONMENTAbstractFog computing has got great attntion due to its importance especially in Internet of Things (IoT) environment where computation at the edge of the network is most desired. Due to the geographical proximity of resources, Fog computing exhibits lower latency compared to cloud; however, inefficient resource allocation in Fog environment can result in higher delays and degraded performance. Hence, efficient resource scheduling in Fog computing is crucial to get true benefits of the cloud like services at the proximity of data generation sources. In this paper, a Big-Medium-Little (BML) scheduling technique is proposed to efficiently allocate Fog and Cloud resources to the incoming IoT jobs. Moreover, cooperative and non-cooperative Fog computing environments are also explored. Additionally, a thorough comparative study of existing scheduling techniques in Fog-cloud environment is also presented. The technique is rigorously evaluated and shows promising results in terms of makespan, energy consumption, latecny and throughput.Keywords: Cloud node, Fog node, Max-Min, Min-Min, Big, Medium, Little, Task, Resource, Cooperative and Non-Cooperative Systems

    A hybrid genetic algorithm with mapreduce technique for cloud computing energy efficiency

    Get PDF
    Computer clouds generally comprise large power-consuming data centers as they are designed to support the elasticity and scalability required by customers. However, while cloud computing reduces energy consumption for customers, it is an issue for providers who have to deal with increasing demand and performance expectations. This creates the need for mechanisms to improve the energy-efficiency of cloud computing data centers while maintaining desired levels of performance. This research seeks to formulate a hybrid algorithm based on Genetic algorithm and MapReduce algorithm techniques to further promote energy efficiency in the cloud computing platform. The function of the MapReduce algorithm is to optimize scheduling performance which is one of the more efficient techniques for handling large data in servers. Genetic algorithm is effective in optimally measuring the value of operations and allows for the minimization of energy efficiency where it includes the formula for single optimization energy efficiency. A series of simulations were developed to evaluate the effectiveness of the proposed algorithm. The evaluation results show the amount of Information Technology equipment power required for Power Usage Effectiveness values to optimize energy usage where the performance of the proposed algorithm is 6% better than the previous genetic algorithm and 5% better than Hadoop MapReduce scheduling on low load conditions. On the other hand, the proposed algorithm improved energy efficiency in comparison with the previous work

    A requirement-driven mechanism for the management of distributed infrastructures

    Get PDF

    Efficient Load Balancing Algorithm in Cloud Computing

    Get PDF
    Recently, cloud computing become a new global trend of computing. It is a modern style of using the power of Internet and wide area network (WAN) to offer resources remotely. It’s a new solution and strategy to achieve high availability, flexibility, cost reduced and on demand scalability. However cloud computing has many challenges such as poor resource utilization which has deep impact in the performance of cloud computing. These problems arisen due to the huge amounts of information. So the need for efficient and powerful cloud computing load balancing algorithms is one of the most important issues in this area to improve the performance of cloud computing. Many researchers proposed various load balancing and job scheduling algorithms in cloud computing, but there is still some inefficiency in the system performance and load still imbalance. Therefore, in this research we propose a load balancing algorithm to improve the performance and efficiency in heterogeneous cloud computing environment. We propose a hybrid algorithm based on randomization and greedy algorithm, it takes advantages of both random and greedy algorithms. The algorithm considers the current resource information and the CPU capacity factor to achieve the objectives. The hybrid algorithm has been evaluated and compared with other algorithms using CloudAnalyst simulator. The results showed improvements on average response time and on processing time by considering the current resource information and the CPU capacity factor compared with other algorithms, and this means the performance has improved

    Secure Data Management and Transmission Infrastructure for the Future Smart Grid

    Get PDF
    Power grid has played a crucial role since its inception in the Industrial Age. It has evolved from a wide network supplying energy for incorporated multiple areas to the largest cyber-physical system. Its security and reliability are crucial to any country’s economy and stability [1]. With the emergence of the new technologies and the growing pressure of the global warming, the aging power grid can no longer meet the requirements of the modern industry, which leads to the proposal of ‘smart grid’. In smart grid, both electricity and control information communicate in a massively distributed power network. It is essential for smart grid to deliver real-time data by communication network. By using smart meter, AMI can measure energy consumption, monitor loads, collect data and forward information to collectors. Smart grid is an intelligent network consists of many technologies in not only power but also information, telecommunications and control. The most famous structure of smart grid is the three-layer structure. It divides smart grid into three different layers, each layer has its own duty. All these three layers work together, providing us a smart grid that monitor and optimize the operations of all functional units from power generation to all the end-customers [2]. To enhance the security level of future smart grid, deploying a high secure level data transmission scheme on critical nodes is an effective and practical approach. A critical node is a communication node in a cyber-physical network which can be developed to meet certain requirements. It also has firewalls and capability of intrusion detection, so it is useful for a time-critical network system, in other words, it is suitable for future smart grid. The deployment of such a scheme can be tricky regarding to different network topologies. A simple and general way is to install it on every node in the network, that is to say all nodes in this network are critical nodes, but this way takes time, energy and money. Obviously, it is not the best way to do so. Thus, we propose a multi-objective evolutionary algorithm for the searching of critical nodes. A new scheme should be proposed for smart grid. Also, an optimal planning in power grid for embedding large system can effectively ensure every power station and substation to operate safely and detect anomalies in time. Using such a new method is a reliable method to meet increasing security challenges. The evolutionary frame helps in getting optimum without calculating the gradient of the objective function. In the meanwhile, a means of decomposition is useful for exploring solutions evenly in decision space. Furthermore, constraints handling technologies can place critical nodes on optimal locations so as to enhance system security even with several constraints of limited resources and/or hardware. The high-quality experimental results have validated the efficiency and applicability of the proposed approach. It has good reason to believe that the new algorithm has a promising space over the real-world multi-objective optimization problems extracted from power grid security domain. In this thesis, a cloud-based information infrastructure is proposed to deal with the big data storage and computation problems for the future smart grid, some challenges and limitations are addressed, and a new secure data management and transmission strategy regarding increasing security challenges of future smart grid are given as well
    corecore