5,740 research outputs found
Hybrid ant colony system and genetic algorithm approach for scheduling of jobs in computational grid
Metaheuristic algorithms have been used to solve scheduling problems in grid computing.However,
stand-alone metaheuristic algorithms do not always show good performance in every problem instance. This study proposes a high level hybrid approach between ant colony system and genetic algorithm for job scheduling in grid computing.The proposed approach is based on a high level hybridization.The proposed hybrid approach is evaluated using the static benchmark problems known as ETC matrix.Experimental results show that the proposed hybridization between the two algorithms outperforms the stand-alone algorithms in terms of best and average
makespan values
Applying autonomy to distributed satellite systems: Trends, challenges, and future prospects
While monolithic satellite missions still pose significant advantages in terms of accuracy and
operations, novel distributed architectures are promising improved flexibility, responsiveness,
and adaptability to structural and functional changes. Large satellite swarms, opportunistic satellite
networks or heterogeneous constellations hybridizing small-spacecraft nodes with highperformance
satellites are becoming feasible and advantageous alternatives requiring the adoption
of new operation paradigms that enhance their autonomy. While autonomy is a notion that
is gaining acceptance in monolithic satellite missions, it can also be deemed an integral characteristic
in Distributed Satellite Systems (DSS). In this context, this paper focuses on the motivations
for system-level autonomy in DSS and justifies its need as an enabler of system qualities. Autonomy
is also presented as a necessary feature to bring new distributed Earth observation functions
(which require coordination and collaboration mechanisms) and to allow for novel structural
functions (e.g., opportunistic coalitions, exchange of resources, or in-orbit data services). Mission
Planning and Scheduling (MPS) frameworks are then presented as a key component to implement
autonomous operations in satellite missions. An exhaustive knowledge classification explores the
design aspects of MPS for DSS, and conceptually groups them into: components and organizational
paradigms; problem modeling and representation; optimization techniques and metaheuristics;
execution and runtime characteristics and the notions of tasks, resources, and constraints.
This paper concludes by proposing future strands of work devoted to study the trade-offs of
autonomy in large-scale, highly dynamic and heterogeneous networks through frameworks that
consider some of the limitations of small spacecraft technologies.Postprint (author's final draft
Scientific Workflow Scheduling for Cloud Computing Environments
The scheduling of workflow applications consists of assigning their tasks to computer resources to fulfill a final goal such as minimizing total workflow execution time. For this reason, workflow scheduling plays a crucial role in efficiently running experiments. Workflows often have many discrete tasks and the number of different task distributions possible and consequent time required to evaluate each configuration quickly becomes prohibitively large. A proper solution to the scheduling problem requires the analysis of tasks and resources, production of an accurate environment model and, most importantly, the adaptation of optimization techniques. This study is a major step toward solving the scheduling problem by not only addressing these issues but also optimizing the runtime and reducing monetary cost, two of the most important variables. This study proposes three scheduling algorithms capable of answering key issues to solve the scheduling problem. Firstly, it unveils BaRRS, a scheduling solution that exploits parallelism and optimizes runtime and monetary cost. Secondly, it proposes GA-ETI, a scheduler capable of returning the number of resources that a given workflow requires for execution. Finally, it describes PSO-DS, a scheduler based on particle swarm optimization to efficiently schedule large workflows. To test the algorithms, five well-known benchmarks are selected that represent different scientific applications. The experiments found the novel algorithms solutions substantially improve efficiency, reducing makespan by 11% to 78%. The proposed frameworks open a path for building a complete system that encompasses the capabilities of a workflow manager, scheduler, and a cloud resource broker in order to offer scientists a single tool to run computationally intensive applications
Hybrid ant colony system algorithm for static and dynamic job scheduling in grid computing
Grid computing is a distributed system with heterogeneous infrastructures. Resource
management system (RMS) is one of the most important components which has great influence on the grid computing performance. The main part of RMS is the scheduler algorithm which has the responsibility to map submitted tasks to available resources. The complexity of scheduling problem is considered as a nondeterministic polynomial complete (NP-complete) problem and therefore, an intelligent algorithm is required to achieve better scheduling solution. One of the prominent intelligent algorithms is ant colony system (ACS) which is implemented widely to solve various types of scheduling problems. However, ACS suffers from stagnation problem in medium and large size grid computing system. ACS is based on exploitation and exploration
mechanisms where the exploitation is sufficient but the exploration has a deficiency. The exploration in ACS is based on a random approach without any strategy. This study proposed four hybrid algorithms between ACS, Genetic Algorithm (GA), and Tabu Search (TS) algorithms to enhance the ACS performance. The algorithms are ACS(GA), ACS+GA, ACS(TS), and ACS+TS. These proposed hybrid algorithms
will enhance ACS in terms of exploration mechanism and solution refinement by
implementing low and high levels hybridization of ACS, GA, and TS algorithms. The proposed algorithms were evaluated against twelve metaheuristic algorithms in static (expected time to compute model) and dynamic (distribution pattern) grid computing
environments. A simulator called ExSim was developed to mimic the static and dynamic nature of the grid computing. Experimental results show that the proposed algorithms outperform ACS in terms of best makespan values. Performance of ACS(GA), ACS+GA, ACS(TS), and ACS+TS are better than ACS by 0.35%, 2.03%, 4.65% and 6.99% respectively for static environment. For dynamic environment,
performance of ACS(GA), ACS+GA, ACS+TS, and ACS(TS) are better than ACS by 0.01%, 0.56%, 1.16%, and 1.26% respectively. The proposed algorithms can be used to schedule tasks in grid computing with better performance in terms of makespan
A Time-driven Data Placement Strategy for a Scientific Workflow Combining Edge Computing and Cloud Computing
Compared to traditional distributed computing environments such as grids,
cloud computing provides a more cost-effective way to deploy scientific
workflows. Each task of a scientific workflow requires several large datasets
that are located in different datacenters from the cloud computing environment,
resulting in serious data transmission delays. Edge computing reduces the data
transmission delays and supports the fixed storing manner for scientific
workflow private datasets, but there is a bottleneck in its storage capacity.
It is a challenge to combine the advantages of both edge computing and cloud
computing to rationalize the data placement of scientific workflow, and
optimize the data transmission time across different datacenters. Traditional
data placement strategies maintain load balancing with a given number of
datacenters, which results in a large data transmission time. In this study, a
self-adaptive discrete particle swarm optimization algorithm with genetic
algorithm operators (GA-DPSO) was proposed to optimize the data transmission
time when placing data for a scientific workflow. This approach considered the
characteristics of data placement combining edge computing and cloud computing.
In addition, it considered the impact factors impacting transmission delay,
such as the band-width between datacenters, the number of edge datacenters, and
the storage capacity of edge datacenters. The crossover operator and mutation
operator of the genetic algorithm were adopted to avoid the premature
convergence of the traditional particle swarm optimization algorithm, which
enhanced the diversity of population evolution and effectively reduced the data
transmission time. The experimental results show that the data placement
strategy based on GA-DPSO can effectively reduce the data transmission time
during workflow execution combining edge computing and cloud computing
An Efficient Firefly Algorithm for Optimizing Task Scheduling in Cloud Computing Systems
As user service demands change constantly, task scheduling becomes an extremely significant study area within the cloud environment. The goal of scheduling is distributing the tasks on available processors in order to achieve the shortest possible makespan while adhering to priority constraints. In heterogeneous cloud computing resources, task scheduling has a large influence on system performances. The various processes in the heuristic-based algorithm of scheduling will result in varied makespans when heterogeneous resources are utilized. As a result, a smart method of scheduling must be capable of establishing precedence efficacy for each task to decrease makespan time. In our study, we develop a novel efficient method of scheduling tasks according to the firefly algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem. We evaluate the performance of our algorithm by putting it through three situations with changing amounts of processors and numbers of tasks. The findings of the experiment reveal that our suggested technique found optimal solutions substantially more frequently in terms of makespan time when compared with other methods
Low and high level hybridization of ant colony system and genetic algorithm for job scheduling in grid computing
Hybrid metaheuristic algorithms have the ability to produce better solution than stand-alone approach and no algorithm could be concluded as the best algorithm for scheduling algorithm or in general, for combinatorial problems.This study presents the low and high level hybridization of ant colony system and genetic algorithm in solving the job scheduling in grid computing.Two hybrid algorithms namely ACS(GA) as a low level and ACS+GA as a high level are proposed.The proposed algorithms were evaluated using static benchmarks problems known as expected time to compute model. Experimental results show that ant colony system algorithm performance is enhanced when hybridized with genetic algorithm specifically with high level hybridization
Hybrid Meta-heuristic Algorithms for Static and Dynamic Job Scheduling in Grid Computing
The term ’grid computing’ is used to describe an infrastructure that connects geographically
distributed computers and heterogeneous platforms owned by multiple organizations
allowing their computational power, storage capabilities and other resources to be selected
and shared. Allocating jobs to computational grid resources in an efficient manner is one
of the main challenges facing any grid computing system; this allocation is called job
scheduling in grid computing. This thesis studies the application of hybrid meta-heuristics
to the job scheduling problem in grid computing, which is recognized as being one of
the most important and challenging issues in grid computing environments. Similar to
job scheduling in traditional computing systems, this allocation is known to be an NPhard
problem. Meta-heuristic approaches such as the Genetic Algorithm (GA), Variable
Neighbourhood Search (VNS) and Ant Colony Optimisation (ACO) have all proven their
effectiveness in solving different scheduling problems. However, hybridising two or more
meta-heuristics shows better performance than applying a stand-alone approach. The new
high level meta-heuristic will inherit the best features of the hybridised algorithms, increasing
the chances of skipping away from local minima, and hence enhancing the overall
performance. In this thesis, the application of VNS for the job scheduling problem in grid
computing is introduced. Four new neighbourhood structures, together with a modified
local search, are proposed. The proposed VNS is hybridised using two meta-heuristic
methods, namely GA and ACO, in loosely and strongly coupled fashions, yielding four
new sequential hybrid meta-heuristic algorithms for the problem of static and dynamic
single-objective independent batch job scheduling in grid computing. For the static version
of the problem, several experiments were carried out to analyse the performance of the
proposed schedulers in terms of minimising the makespan using well known benchmarks.
The experiments show that the proposed schedulers achieved impressive results compared
to other traditional, heuristic and meta-heuristic approaches selected from the bibliography.
To model the dynamic version of the problem, a simple simulator, which uses
the rescheduling technique, is designed and new problem instances are generated, by
using a well-known methodology, to evaluate the performance of the proposed hybrid
schedulers. The experimental results show that the use of rescheduling provides significant
improvements in terms of the makespan compared to other non-rescheduling approaches
- …