11 research outputs found

    Max-Average: An Extended Max-Min Scheduling Algorithm for Grid Computing Environtment

    Get PDF
    Sharing numerous computational and communication power from connected heterogeneous systems over the world are the two key points of Grid computing. Grid computing can also be referred as a computing platform for users to utilise the remote heterogeneous resources for solving their large scale jobs that require a huge amount of processing power or a huge data storage. Sharing these resources that way effectively requires a very good scheduling strategy, which is the focus of this research. This paper presents a new proposed grid based scheduling algorithm called Max-Average, inspired from Max-Min algorithm. In order to produce good quality solutions, the proposed algorithm is designed in two phases; firstly it uses an initial task queue like the traditional Max -Min for estimating task completion time for each of resources, and in the second phase choose the fitting resource for scheduling according to requirements. The results from our simulation showed that our proposed algorithm is performing better in producing good quality solutions, particularly in executing tasks fast and in balancing the load (resource utilisation) among the resources more effectively when compared to standard Minimum Execution Time (MET), Minimum Completion Time (MCT), Min-Min, and Max-Min heuristic approache

    An ordered heuristic for the allocation of resources in unrelated parallel-machines

    Get PDF
    All rights reserved. Global competition pressures have forced manufactures to adapt their productive capabilities. In order to satisfy the ever-changing market demands many organizations adopted flexible resources capable of executing several products with different performance criteria. The unrelated parallel-machines makespan minimization problem (Rm||Cmax) is known to be NP-hard or too complex to be solved exactly. In the heuristics used for this problem, the MCT (Minimum Completion Time), which is the base for several others, allocates tasks in a random like order to the minimum completion time machine. This paper proposes an ordered approach to the MCT heuristic. MOMCT (Modified Ordered Minimum Completion Time) will order tasks in accordance to the MS index, which represents the mean difference of the completion time on each machine and the one on the minimum completion time machine. The computational study demonstrates the improved performance of MOMCT over the MCT heuristic.This work is supported by FEDER Funds through the “Programa Operacional Factores de Competitividade - COMPETE” program and by National Funds through FCT “Fundação para a Ciência e a Tecnologia” under the project: FCOMP-01-0124-FEDER-PEst-OE/EEI/UI0760/2011 and PEstOE/EEI/UI0760/2014.info:eu-repo/semantics/publishedVersio

    IMMEDIATE/BATCH MODE SCHEDULING ALGORITHMS FOR GRID COMPUTING: A REVIEW

    Get PDF
    Immediate/on-line and Batch mode heuristics are two methods used for scheduling in the computational grid environment. In the former, task is mapped onto a resource as soon as it arrives at the scheduler, while the later, tasks are not mapped onto resource as they arrive, instead they are collected into a set that is examined for mapping at prescheduled times called mapping events. This paper reviews the literature concerning Minimum Execution Time (MET) along with Minimum Completion Time (MCT) algorithms of online mode heuristics and more emphasis on Min-Min along with Max-Min algorithms of batch mode heuristics, while focusing on the details of their basic concepts, approaches, techniques, and open problems

    A Novel Decentralized Fuzzy Based Approach for Grid Job

    Get PDF
    In this paper with the aid of fuzzy theory we present a new method for scheduling on Grid system. Grid computing is a technology to meet the growing computational requires. In fact grid computing is one of the most popular types of distributed system. Its aim is to produce an enormous, autonomous and effective virtual machine, and it is produced by collecting different nodes with the aim of sharing their data and computational power. This paper follows the identification of grid scheduling with the help of fuzzy theory and seeking to present a new method for grid scheduling with respect to exiting obstacles. In our method we use the intermediate load of nodes of each clusters, the average of computing power which determines the node premiership and job premiership as the input parameters of fuzzy system, and regarding to the output value of fuzzy system the suitable nodes determines. We evaluate the performance of our method with some grid scheduling methods. The results of the experiments show the efficiency of the proposed method in term of makespan and Standard deviation of the load of cluster

    A P2P Computing System for Overlay Networks

    Full text link
    A distributed computing system is able to perform data computation and distribution of results at the same time. The input task is divided into blocks, which are then sent to system participants that offer their resources in order to perform calculations. Next, a partial result is sent back by the participants to the task manager (usually one central node). In the case when system participants want to get the final result, the central node may become overloaded, especially if many nodes request the result at the same time. In this paper we propose a novel distributed computation system, which does not use the central node as the source of the final result, but assumes that partial results are sent between system participants. This way we avoid overloading the central node, as well as network congestion. There are two major types of distributed computing systems: grids and Peer-to-Peer (P2P) computing systems. In this work we focus on the latter case. Consequently, we assume that the computing system works on the top of an overlay network. We present a complete description of the P2P computing system, considering both computation and result distribution. To verify the proposed architecture we develop our own simulator. The obtained results show the system performance expressed by the operation cost for various types of network flows: unicast, anycast and Peer-to-Peer. Moreover, the simulations prove that our computing system provides about 66% lower cost compared to a centralized computing system

    Decision Strategies for a P2P Computing System

    Full text link
    Peer-to-Peer (P2P) computing (also called ‘public-resource computing’) is an effective approach to perform computation of large tasks. Currently used P2P computing systems (e.g., BOINC) are most often centrally managed, i.e., the final result of computations is created at a central node using partial results – what may be not efficient in the case when numerous participants are willing to download the final result. In this paper, we propose a novel approach to P2P computing systems. We assume that results can be delivered to all peers in a distributed way using three types of network flows: unicast, Peer-to-Peer and anycast. We describe our concept of the system architecture with a special focus on the decision strategies developed for system participants. Using our discrete realtime simulator we evaluate the proposed strategies in various scenarios and present a comprehensive analysis of obtained results. According to obtained results, the Peer-to-Peer flow provides lower operational cost of the computing system compared to unicast and anycast flows. Moreover, an appropriate selection of decision strategy can significantly reduce the operational cost

    A Comparison among Grid Scheduling Algorithms for Independent Coarse-Grained Tasks

    No full text
    The most common objective function of task scheduling problems is makespan. However, on a computational grid, the 2nd optimal makespan may be much longer than the optimal makespan because the computing power of a grid varies over time. So, if the performance measure is makespan, there is no approximation algorithm in general for scheduling onto a grid. In contrast, recently the authors proposed the computing power consumed by a schedule as a criterion of the schedule and, for the criterion, gave (1 + )-approximation algorithm RR for scheduling n independent coarse-grained tasks with the same length onto a grid with m processors. RR does not use any prediction information on the underlying resources. RR is the first approximation algorithm for grid scheduling. However, so far any performance comparison among related heuristic algorithms is not given. This paper shows experimental results on the comparison of the consumed computing power of a schedule among RR and five related algorithms. It turns out that RR is next to the best of algorithms that need the prediction information on processor speeds and task lengths though RR does not require such information

    Enhancing reliability with Latin Square redundancy on desktop grids.

    Get PDF
    Computational grids are some of the largest computer systems in existence today. Unfortunately they are also, in many cases, the least reliable. This research examines the use of redundancy with permutation as a method of improving reliability in computational grid applications. Three primary avenues are explored - development of a new redundancy model, the Replication and Permutation Paradigm (RPP) for computational grids, development of grid simulation software for testing RPP against other redundancy methods and, finally, running a program on a live grid using RPP. An important part of RPP involves distributing data and tasks across the grid in Latin Square fashion. Two theorems and subsequent proofs regarding Latin Squares are developed. The theorems describe the changing position of symbols between the rows of a standard Latin Square. When a symbol is missing because a column is removed the theorems provide a basis for determining the next row and column where the missing symbol can be found. Interesting in their own right, the theorems have implications for redundancy. In terms of the redundancy model, the theorems allow one to state the maximum makespan in the face of missing computational hosts when using Latin Square redundancy. The simulator software was developed and used to compare different data and task distribution schemes on a simulated grid. The software clearly showed the advantage of running RPP, which resulted in faster completion times in the face of computational host failures. The Latin Square method also fails gracefully in that jobs complete with massive node failure while increasing makespan. Finally an Inductive Logic Program (ILP) for pharmacophore search was executed, using a Latin Square redundancy methodology, on a Condor grid in the Dahlem Lab at the University of Louisville Speed School of Engineering. All jobs completed, even in the face of large numbers of randomly generated computational host failures

    A Globally Distributed System for Job, Data, and Information Handling for High Energy Physics

    Full text link
    corecore