392 research outputs found
Divisible load scheduling of image processing applications on the heterogeneous star and tree networks using a new genetic algorithm
The divisible load scheduling of image processing applications on the heterogeneous star and multi-level tree networks is addressed in this paper. In our platforms, processors and network links have different speeds. In addition, computation and communication overheads are considered. A new genetic algorithm for minimizing the processing time of low-level image applications using divisible load theory is introduced. The closed-form solution for the processing time, the image fractions that should be allocated to each processor, the optimum number of participating processors, and the optimal sequence for load distribution are derived. The new concept of equivalent processor in tree network is introduced and the effect of different image and kernel sizes on processing time and speed up are investigated. Finally, to indicate the efficiency of our algorithm, several numerical experiments are presented
Design and analysis of load balancing/scheduling strategies on distributed computer networks using virtual routing approach
Ph.DDOCTOR OF PHILOSOPH
Recommended from our members
Personal mobile grids with a honeybee inspired resource scheduler
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The overall aim of the thesis has been to introduce Personal Mobile Grids (PMGrids)
as a novel paradigm in grid computing that scales grid infrastructures to mobile devices and extends grid entities to individual personal users. In this thesis, architectural designs as well as simulation models for PM-Grids are developed.
The core of any grid system is its resource scheduler. However, virtually all current conventional grid schedulers do not address the non-clairvoyant scheduling problem, where job information is not available before the end of execution. Therefore, this thesis proposes a honeybee inspired resource scheduling heuristic for PM-Grids (HoPe) incorporating a radical approach to grid resource scheduling to tackle this problem. A detailed design and implementation of HoPe with a decentralised self-management and adaptive policy are initiated.
Among the other main contributions are a comprehensive taxonomy of grid systems as well as a detailed analysis of the honeybee colony and its nectar acquisition process (NAP), from the resource scheduling perspective, which have not been presented in any previous work, to the best of our knowledge.
PM-Grid designs and HoPe implementation were evaluated thoroughly through a strictly controlled empirical evaluation framework with a well-established heuristic in high throughput computing, the opportunistic scheduling heuristic (OSH), as a benchmark algorithm. Comparisons with optimal values and worst bounds are conducted to gain a clear insight into HoPe behaviour, in terms of stability, throughput, turnaround time and speedup, under different running conditions of number of jobs and grid scales.
Experimental results demonstrate the superiority of HoPe performance where it
has successfully maintained optimum stability and throughput in more than 95%
of the experiments, with HoPe achieving three times better than the OSH under
extremely heavy loads. Regarding the turnaround time and speedup, HoPe has
effectively achieved less than 50% of the turnaround time incurred by the OSH, while doubling its speedup in more than 60% of the experiments.
These results indicate the potential of both PM-Grids and HoPe in realising futuristic grid visions. Therefore considering the deployment of PM-Grids in real life scenarios and the utilisation of HoPe in other parallel processing and high throughput computing systems are recommended
Real-time optimization of an integrated production-inventory-distribution problem.
In today\u27s competitive business environment, companies face enormous pressure and must continuously search for ways to design new products, manufacture and distribute them in an efficient and effective fashion. After years of focusing on reduction in production and operation costs, companies are beginning to look into distribution activities as the last frontier for cost reduction. In addition, an increasing number of companies, large and small, are focusing their efforts on their core competencies which are critical to survive. This results in a widespread practice in industry that companies outsource one or more than one logistics functions to third party logistics providers. By using such logistics expertise, they can obtain a competitive advantage both in cost and time efficiency, because the third party logistics companies already have the equipment, system and experience and are ready to help to their best efforts. In this dissertation, we developed an integrated optimization model of production, inventory and distribution with the goal to coordinate important and interrelated decisions related to production schedules, inventory policy and truckload allocation. Because outsourcing logistics functions to third party logistics providers is becoming critical for a company to remain competitive in the market place; we also included an important decision of selecting carriers with finite truckload and drivers for both inbound and outbound shipments in the model. The integrated model is solved by modified Benders decomposition which solves the master problem by a genetic algorithm. Computational results on test problems of various sizes are provided to show the effectiveness of the proposed solution methodology. We also apply this proposed algorithm on a real distribution problem faced by a large national manufacturer and distributor. It shows that such a complex distribution network with 22 plants, 7 distribution centers, 8 customer zones, 9 products, 16 inbound and 16 outbound shipment carriers in a 12-month planning period can be redesigned within 33 hours. In recent years, multi-agent simulation has been a preferred approach to solve logistics and distribution problems, since these problems are autonomous, distributive, complex, heterogeneous and decentralized in nature and they require extensive intelligent decision making. Another important part in this dissertation involved a development of an agent-based simulation model to cooperate with the optimal solution given by the optimization model. More specifically, the solution given by the optimization model can be inputted as the initial condition of the agent-based simulation model. The agent-based simulation model can incorporate many other factors to be considered in the real world, but optimization cannot handle these as needed. The agent-based simulation model can also incorporate some dynamics we may encounter in the real operations, and it can react to these dynamics in real time. Various types of entities in the entire distribution system can be modeled as intelligent agents, such as suppliers, carriers and customers. In order to build the simulation model more realistic, a sealed bid multiunit auction with an introduction of three parameters a, ß and y is well designed. With the help of these three parameters, each agent makes a better decision in a simple and fast manner, which is the key to realizing real-time decision making. After building such a multi-agent system with agent-based simulation approach, it supports more flexible and comprehensive modeling capabilities which are difficult to realize in a general optimization model. The simulation model is tested and validated on an industrial-sized problem. Numerical results of the agent-based simulation model suggest that with appropriate setting of three parameters the model can precisely represent the preference and interest of different decision makers
Distributed Task Management in Cyber-Physical Systems: How to Cooperate under Uncertainty?
We consider the problem of task allocation in a network of cyber-physical
systems (CPSs). The network can have different states, and the tasks are of
different types. The task arrival is stochastic and state-dependent. Every CPS
is capable of performing each type of task with some specific state-dependent
efficiency. The CPSs have to agree on task allocation prior to knowing about
the realized network's state and/or the arrived tasks. We model the problem as
a multi-state stochastic cooperative game with state uncertainty. We then use
the concept of deterministic equivalence and sequential core to solve the
problem. We establish the non-emptiness of the strong sequential core in our
designed task allocation game and investigate its characteristics including
uniqueness and optimality. Moreover, we prove that in the task allocation game,
the strong sequential core is equivalent to Walrasian equilibrium under state
uncertainty; consequently, it can be implemented by using the Walras'
tatonnement process
Scheduling And Resource Management For Complex Systems: From Large-scale Distributed Systems To Very Large Sensor Networks
In this dissertation, we focus on multiple levels of optimized resource management techniques. We first consider a classic resource management problem, namely the scheduling of data-intensive applications. We define the Divisible Load Scheduling (DLS) problem, outline the system model based on the assumption that data staging and all communication with the sites can be done in parallel, and introduce a set of optimal divisible load scheduling algorithms and the related fault-tolerant coordination algorithm. The DLS algorithms introduced in this dissertation exploit parallel communication, consider realistic scenarios regarding the time when heterogeneous computing systems are available, and generate optimal schedules. Performance studies show that these algorithms perform better than divisible load scheduling algorithms based upon sequential communication. We have developed a self-organization model for resource management in distributed systems consisting of a very large number of sites with excess computing capacity. This self-organization model is inspired by biological metaphors and uses the concept of varying energy levels to express activity and goal satisfaction. The model is applied to Pleiades, a service-oriented architecture based on resource virtualization. The self-organization model for complex computing and communication systems is applied to Very Large Sensor Networks (VLSNs). An algorithm for self-organization of anonymous sensor nodes called SFSN (Scale-free Sensor Networks) and an algorithm utilizing the Small-worlds principle called SWAS (Small-worlds of Anonymous Sensors) are introduced. The SFSN algorithm is designed for VLSNs consisting of a fairly large number of inexpensive sensors with limited resources. An important feature of the algorithm is the ability to interconnect sensors without an identity, or physical address used by traditional communication and coordination protocols. During the self-organization phase, the collision-free communication channels allowing a sensor to synchronously forward information to the members of its proximity set are established and the communication pattern is followed during the activity phases. Simulation study shows that the SFSN ensures the scalability, limits the amount of communication and the complexity of coordination. The SWAS algorithm is further improved from SFSN by applying the Small-worlds principle. It is unique in its ability to create a sensor network with a topology approximating small-world networks. Rather than creating shortcuts between pairs of diametrically positioned nodes in a logical ring, we end up with something resembling a double-stranded DNA. By exploiting Small-worlds principle we combine two desirable features of networks, namely high clustering and small path length
Recommended from our members
Computing resources sensitive parallelization of neural neworks for large scale diabetes data modelling, diagnosis and prediction
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Diabetes has become one of the most severe deceases due to an increasing number of diabetes patients globally. A large amount of digital data on diabetes has been collected through various channels. How to utilize these data sets to help doctors to make a decision on diagnosis, treatment and prediction of diabetic patients poses many challenges to the research community. The thesis investigates mathematical models with a focus on neural networks for large scale diabetes data modelling and analysis by utilizing modern computing technologies such as grid computing and cloud computing. These computing technologies provide users with an inexpensive way to have access to extensive computing resources over the Internet for solving data and computationally intensive problems. This thesis evaluates the performance of seven representative machine learning techniques in classification of diabetes data and the results show that neural network produces the best accuracy in classification but incurs high overhead in data training. As a result, the thesis develops MRNN, a parallel neural network model based on the MapReduce programming model which has become an enabling technology in support of data intensive applications in the clouds.
By partitioning the diabetic data set into a number of equally sized data blocks, the workload in training is distributed among a number of computing nodes for speedup in data training. MRNN is first evaluated in small scale experimental environments using 12 mappers and subsequently is evaluated in large scale simulated environments using up to 1000 mappers. Both the experimental and simulations results have shown the effectiveness of MRNN in classification, and its high scalability in data training.
MapReduce does not have a sophisticated job scheduling scheme for heterogonous computing environments in which the computing nodes may have varied computing capabilities. For this purpose, this thesis develops a load balancing scheme based on genetic algorithms with an aim to balance the training workload among heterogeneous computing nodes. The nodes with more computing capacities will receive more MapReduce jobs for execution. Divisible load theory is employed to guide the evolutionary process of the genetic algorithm with an aim to achieve fast convergence. The proposed load balancing scheme is evaluated in large scale simulated MapReduce environments with varied levels of heterogeneity using different sizes of data sets. All the results show that the genetic algorithm based load balancing scheme significantly reduce the makespan in job execution in comparison with the time consumed without load balancing.This work is funded by the EPSRC and China Market Association
- …