29 research outputs found
Optimization Models and Approximate Algorithms for the Aerial Refueling Scheduling and Rescheduling Problems
The Aerial Refueling Scheduling Problem (ARSP) can be defined as determining the refueling completion times for fighter aircrafts (jobs) on multiple tankers (machines) to minimize the total weighted tardiness. ARSP can be modeled as a parallel machine scheduling with release times and due date-to-deadline window. ARSP assumes that the jobs have different release times, due dates, and due date-to-deadline windows between the refueling due date and a deadline to return without refueling. The Aerial Refueling Rescheduling Problem (ARRP), on the other hand, can be defined as updating the existing AR schedule after being disrupted by job related events including the arrival of new aircrafts, departure of an existing aircrafts, and changes in aircraft priorities. ARRP is formulated as a multiobjective optimization problem by minimizing the total weighted tardiness (schedule quality) and schedule instability. Both ARSP and ARRP are formulated as mixed integer programming models. The objective function in ARSP is a piecewise tardiness cost that takes into account due date-to-deadline windows and job priorities. Since ARSP is NP-hard, four approximate algorithms are proposed to obtain solutions in reasonable computational times, namely (1) apparent piecewise tardiness cost with release time rule (APTCR), (2) simulated annealing starting from random solution (SArandom ), (3) SA improving the initial solution constructed by APTCR (SAAPTCR), and (4) Metaheuristic for Randomized Priority Search (MetaRaPS). Additionally, five regeneration and partial repair algorithms (MetaRE, BestINSERT, SEPRE, LSHIFT, and SHUFFLE) were developed for ARRP to update instantly the current schedule at the disruption time. The proposed heuristic algorithms are tested in terms of solution quality and CPU time through computational experiments with randomly generated data to represent AR operations and disruptions. Effectiveness of the scheduling and rescheduling algorithms are compared to optimal solutions for problems with up to 12 jobs and to each other for larger problems with up to 60 jobs. The results show that, APTCR is more likely to outperform SArandom especially when the problem size increases, although it has significantly worse performance than SA in terms of deviation from optimal solution for small size problems. Moreover CPU time performance of APTCR is significantly better than SA in both cases. MetaRaPS is more likely to outperform SAAPTCR in terms of average error from optimal solutions for both small and large size problems. Results for small size problems show that MetaRaPS algorithm is more robust compared to SAAPTCR. However, CPU time performance of SA is significantly better than MetaRaPS in both cases. ARRP experiments were conducted with various values of objective weighting factor for extended analysis. In the job arrival case, MetaRE and BestINSERT have significantly performed better than SEPRE in terms of average relative error for small size problems. In the case of job priority disruption, there is no significant difference between MetaRE, BestINSERT, and SHUFFLE algorithms. MetaRE has significantly performed better than LSHIFT to repair job departure disruptions and significantly superior to the BestINSERT algorithm in terms of both relative error and computational time for large size problems
A common framework and taxonomy for multicriteria scheduling problems with Interfering and competing Jobs: Multi-agent scheduling problems
Most classical scheduling research assumes that the objectives sought are common to all jobs to be
scheduled. However, many real-life applications can be modeled by considering different sets of jobs,
each one with its own objective(s), and an increasing number of papers addressing these problems has
appeared over the last few years. Since so far the area lacks a uni ed view, the studied problems
have received different names (such as interfering jobs, multi-agent scheduling, mixed-criteria, etc), some
authors do not seem to be aware of important contributions in related problems, and solution procedures
are often developed without taking into account existing ones. Therefore, the topic is in need of a common
framework that allows for a systematic recollection of existing contributions, as well as a clear de nition
of the main research avenues. In this paper we review multicriteria scheduling problems involving two or
more sets of jobs and propose an uni ed framework providing a common de nition, name and notation
for these problems. Moreover, we systematically review and classify the existing contributions in terms
of the complexity of the problems and the proposed solution procedures, discuss the main advances, and
point out future research lines in the topic
A DECOMPOSITION-BASED HEURISTIC ALGORITHM FOR PARALLEL BATCH PROCESSING PROBLEM WITH TIME WINDOW CONSTRAINT
This study considers a parallel batch processing problem to minimize the makespan under constraints of arbitrary lot sizes, start time window and incompatible families. We first formulate the problem with a mixed-integer programming model. Due to the NP-hardness of the problem, we develop a decomposition-based heuristic to obtain a near-optimal solution for large-scale problems when computational time is a concern. A two-dimensional saving function is introduced to quantify the value of time and capacity space wasted. Computational experiments show that the proposed heuristic performs well and can deal with large-scale problems efficiently within a reasonable computational time. For the small-size problems, the percentage of achieving optimal solutions by the DH is 94.17%, which indicates that the proposed heuristic is very good in solving small-size problems. For large-scale problems, our proposed heuristic outperforms an existing heuristic from the literature in terms of solution quality
Scheduling Hybrid Flow Lines of Aerospace Composite Manufacturing Systems
Composite manufacturing is a vital part of aerospace manufacturing systems. Applying effective scheduling within these systems can cut the costs in aerospace companies significantly. These systems can be characterized as two-stage Hybrid Flow Shops (HFS) with identical, non-identical and unrelated parallel discrete-processing machines in the first stage and non-identical parallel batch-processing machines in the second stage. The first stage is normally the lay-up process in which the carbon fiber sheets are stacked on the molds (tools). Then, the parts are batched based on the compatibility of their cure recipe before going to the second stage into the autoclave for curing. Autoclaves require enormous capital investment and maximizing their utilization is of utmost importance.
In this thesis, a Mixed Integer Linear Programming (MILP) model is developed to maximize the utilization of the resources in the second stage of this HFS. CPLEX, with an underlying branch and bound algorithm, is used to solve the model. The results show the high level of flexibility and computational efficiency of the proposed model when applied to small and medium-size problems. However, due to the NP-hardness of the problem, the MILP model fails to solve large problems (i.e. problems with more than 120 jobs as input) in reasonable CPU times.
To solve the larger instances of the problem, a novel heuristic method along with a Genetic Algorithm (GA) are developed. The heuristic algorithm is designed based on a careful observation of the behavior of the MILP model for different problem sets. Moreover, it is enhanced by adding a number of proper dispatching rules. As its output, this heuristic algorithm generates eight initial feasible solutions which are then used as the initial population of the proposed GA.
The GA improves the initial solutions obtained from the aforementioned heuristic through its stochastic iterations until it reaches the satisfactory near-optimal solutions. A novel crossover operator is introduced in this GA which is unique to the HFS of aerospace composite manufacturing systems. The proposed GA is proven to be very efficient when applied to large-size problems with up to 300 jobs. The results show the high quality of the solutions achieved by the GA when compared to the optimal solutions which are obtained from the MILP model.
A real case study undertaken at one of the leading companies in the Canadian aerospace industry is used for the purpose of data experiments and analysis
Recommended from our members
An intelligent manufacturing system for heat treatment scheduling
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.This research is focused on the integration problem of process planning and scheduling in steel heat treatment operations environment using artificial intelligent techniques that are capable of dealing with such problems.
This work addresses the issues involved in developing a suitable methodology for scheduling heat treatment operations of steel. Several intelligent algorithms have been developed for these propose namely, Genetic Algorithm (GA), Sexual Genetic Algorithm (SGA), Genetic Algorithm with Chromosome differentiation (GACD), Age Genetic Algorithm (AGA), and Mimetic Genetic Algorithm (MGA). These algorithms have been employed to develop an efficient intelligent algorithm using Algorithm Portfolio methodology. After that all the algorithms have been tested on two types of scheduling benchmarks.
To apply these algorithms on heat treatment scheduling, a furnace model is developed for optimisation proposes. Furthermore, a system that is capable of selecting the optimal heat treatment regime is developed so the required metal properties can be achieved with the least energy consumption and the shortest time using Neuro-Fuzzy (NF) and Particle Swarm Optimisation (PSO) methodologies. Based on this system, PSO is used to optimise the heat treatment process by selecting different heat treatment conditions. The selected conditions are evaluated so the best selection can be identified. This work addresses the issues involved in developing a suitable methodology for developing an NF system and PSO for mechanical properties of the steel.
Using the optimisers, furnace model and heat treatment system model, the intelligent system model is developed and implemented successfully. The results of this system were exciting and the optimisers were working correctly
Approximate Algorithms for the Combined arrival-Departure Aircraft Sequencing and Reactive Scheduling Problems on Multiple Runways
The problem addressed in this dissertation is the Aircraft Sequencing Problem (ASP) in which a schedule must be developed to determine the assignment of each aircraft to a runway, the appropriate sequence of aircraft on each runway, and their departing or landing times. The dissertation examines the ASP over multiple runways, under mixed mode operations with the objective of minimizing the total weighted tardiness of aircraft landings and departures simultaneously. To prevent the dangers associated with wake-vortex effects, separation times enforced by Aviation Administrations (e.g., FAA) are considered, adding another level of complexity given that such times are sequence-dependent. Due to the problem being NP-hard, it is computationally difficult to solve large scale instances in a reasonable amount of time. Therefore, three greedy algorithms, namely the Adapted Apparent Tardiness Cost with Separation and Ready Times (AATCSR), the Earliest Ready Time (ERT) and the Fast Priority Index (FPI) are proposed. Moreover, metaheuristics including Simulated Annealing (SA) and the Metaheuristic for Randomized Priority Search (Meta-RaPS) are introduced to improve solutions initially constructed by the proposed greedy algorithms. The performance (solution quality and computational time) of the various algorithms is compared to the optimal solutions and to each other.
The dissertation also addresses the Aircraft Reactive Scheduling Problem (ARSP) as air traffic systems frequently encounter various disruptions due to unexpected events such as inclement weather, aircraft failures or personnel shortages rendering the initial plan suboptimal or even obsolete in some cases. This research considers disruptions including the arrival of new aircraft, flight cancellations and aircraft delays. ARSP is formulated as a multi-objective optimization problem in which both the schedule\u27s quality and stability are of interest. The objectives consist of the total weighted start times (solution quality), total weighted start time deviation, and total weighted runway deviation (instability measures). Repair and complete regeneration approximate algorithms are developed for each type of disruptive events. The algorithms are tested against difficult benchmark problems and the solutions are compared to optimal solutions in terms of solution quality, schedule stability and computational time
An intelligent manufacturing system for heat treatment scheduling
This research is focused on the integration problem of process planning and scheduling in steel heat treatment operations environment using artificial intelligent techniques that are capable of dealing with such problems. This work addresses the issues involved in developing a suitable methodology for scheduling heat treatment operations of steel. Several intelligent algorithms have been developed for these propose namely, Genetic Algorithm (GA), Sexual Genetic Algorithm (SGA), Genetic Algorithm with Chromosome differentiation (GACD), Age Genetic Algorithm (AGA), and Mimetic Genetic Algorithm (MGA). These algorithms have been employed to develop an efficient intelligent algorithm using Algorithm Portfolio methodology. After that all the algorithms have been tested on two types of scheduling benchmarks. To apply these algorithms on heat treatment scheduling, a furnace model is developed for optimisation proposes. Furthermore, a system that is capable of selecting the optimal heat treatment regime is developed so the required metal properties can be achieved with the least energy consumption and the shortest time using Neuro-Fuzzy (NF) and Particle Swarm Optimisation (PSO) methodologies. Based on this system, PSO is used to optimise the heat treatment process by selecting different heat treatment conditions. The selected conditions are evaluated so the best selection can be identified. This work addresses the issues involved in developing a suitable methodology for developing an NF system and PSO for mechanical properties of the steel. Using the optimisers, furnace model and heat treatment system model, the intelligent system model is developed and implemented successfully. The results of this system were exciting and the optimisers were working correctly.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
Recommended from our members
Bi-Criteria Batching and Scheduling in Hybrid Flow Shops
In this research, a bi-criteria batching and scheduling problem is investigated in hybrid flow shop environments, where unrelated-parallel machines are run simultaneously with different capacities and eligibilities in processing, in some stages. The objective is to simultaneously minimize a linear combination of the total weighted completion time and total weighted tardiness. The first favors the producer’s interest by minimizing work-in-process inventory, inventory holding cost, and energy consumption as well as maximizing machine utilization, while the second favors the customers’ interest by maximizing customers’ service level and delivery speed. In particular, it disregards the group technology assumptions (GTAs) by allowing for the possibility of splitting pre-determined groups of jobs into inconsistent batches in order to improve the operational efficiency. A comparison between the group scheduling and batch scheduling approaches reveals the outstanding performance of the batch scheduling approach. As a result, contrary to the GTAs, jobs belonging to a group might be processed on more than one machine as batches, but not all machines may be capable of processing all jobs. A sequence- and machine-dependent setup time is required between each of two consecutively scheduled batches belonging to different groups. Based on manufacturing company policy, the desired lower bounds on batch sizes are considered for the number of jobs assigned to batches. Although, the direction in which all jobs move through production line is the same, some jobs may skip some stages. Furthermore, to reflect real industry requirements, the job release times and the machine availability times are considered to be dynamic, which means not all machines and jobs are available at the beginning of the planning horizon.The problem is formulated with the help of four mixed-integer linear programming (MILP) models. Two out of four MILP models are formulated as two integrated phases, i.e., batching and scheduling phases, with respect to the precedence constraints between each pair of jobs batches and or the position concept within batches. The optimal combination between batch compositions of groups are determined in the batching phase, while the optimal assignment and sequence of batches on machines and sequence of jobs within batches are determined in the scheduling phase, with respect to a set of operational constraints. A batch composition of a group corresponding to a particular stage, determined in the batching phase of the MILP model, represents the number of batches assigned to the group as well as the number and type of jobs belonging to each batch of that group. Since the first and second MILP models lead to unmanageable solution space, the relaxed MILP model, which allocates one and only one job to each batch of each group in each stage, can be developed to focus on the non-dominated solution space. The optimal solutions of MILP models and relaxed MILP model are equal, if and only if the optimal solution of the relaxed MILP model does not violate the desired lower bounds on batch sizes. Since the relaxed MILP model cannot guarantee the optimal solution of the MILP models, a third MILP model is developed by integrating batching and scheduling phases. This MILP model eliminates an exhaustive combination enumeration between batch compositions of all groups in all stages. Although the third MILP model converges to the optimal solution slower than the relaxed MILP model, it guarantees finding the optimal solution of the first and second MILP models. A comparison between four MILP models shows the superior performance of the third MILP model. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases from small to medium to large, even by the relaxed MILP model or the fourth MILP model. Therefore, several meta-heuristic algorithms based upon basic local search, basic population-based search, and hybridization of local search and population-based searches are developed, which move back and forth between batching and scheduling phases. Tabu Search (TS) is implemented as a basic local search algorithm, while Tabu Search Path-Relinking (TS PR) is implemented as a local search algorithm enhanced with a population-based structure. TS is incorporated into the framework of path-relinking to exploit the information on good solutions. The TS PR algorithm comprises several distinguishing features including relinking procedures to effectively explore trajectories connecting elite solutions and the methods used to choose the reference solution. Particle Swarm Optimization (PSO) is implemented as a basic population-based algorithm, while Particle Swarm Optimization enhanced with a local search algorithm (PSO LSA) is developed to realize the benefits of batching and, consequently, enhance the quality of solutions.Since there is interdependency between positions of a job in different stages of a hybrid flow shop in batch scheduling, a meta-heuristic algorithm is not capable of capturing these interdependencies and, subsequently, its efficacy can be diminished. In order to capture this interdependency, the non-, partial- complete-, and stage-based interdependency strategy are developed. In the stage-based-interdependency strategy, a complete sequence related to all of the stages is gradually determined, stage by stage. An initial solution finding mechanism is developed to trigger the search into the solution space and generate an initial population. The performances of these algorithms are compared to each other in order to identify which algorithm(s) outperforms the others. Nevertheless, the performances of the best algorithm(s) are evaluated with respect to a tight lower bound obtained from a branch-and-price (B&P) algorithm. The B&P algorithm uses Dantzig-Wolfe decomposition (DWD) to divide the original problem into a master problem and several sub-problems (SPs) corresponding to each stage. The original problem is decomposed into the SPs by three DWDs corresponding to the three MILP models. Although, by applying DWD technique in the first and second MILP models, an exhaustive combination enumeration between batch compositions of all groups in all stages is excluded and, as a result, the SPs are easier to solve than the original problem, they are still strongly NP-hard because of an enormous number of combinations between batch compositions of all groups in each stage. However, the DWD technique corresponding to the relaxed MILP model not only drastically reduces the number of variables and constraints in the SPs, but also eliminates the batching phase of the first and second MILP models. Decomposing the original problem based on the relaxed MILP model and implementing the B&P algorithm cannot guarantee optimal solutions or tight lower bounds of problems unless the number of violations in the desired lower bounds on batch sizes is not significant. Therefore, the third MILP model is decomposed by DWD so that the B&P algorithm is capable of finding tight lower bounds even for large-size instances of the problem. A comparison between the lower bounds obtained from the B&P algorithm and CPLEX reveals the impressive performance of the B&P algorithm, particularly for large-size problems. The evaluation of the best algorithms based upon these tight lower bounds developed by the B&P algorithm, uncovers the outstanding performance of hybrid algorithms compared to the results obtained from CPLEX.Keywords: Bi-Criteria Objective, Column Generation, Batch Scheduling, Tabu Search, Batching and Scheduling, Desired Lower Bounds on Batch Sizes, Path-Relinking, Branch-and-Price Optimization Algorithm, Particle Swarm Optimization, Group Scheduling, Hybrid Flow Shop, Dantzig-Wolfe Decomposition, Mixed-Integer Linear Programming Model, Sequence- and Machine-Dependent Setup Tim
Recommended from our members
Bi-Criteria Batching and Scheduling in Hybrid Flow Shops
In this research, a bi-criteria batching and scheduling problem is investigated in hybrid flow shop environments, where unrelated-parallel machines are run simultaneously with different capacities and eligibilities in processing, in some stages. The objective is to simultaneously minimize a linear combination of the total weighted completion time and total weighted tardiness. The first favors the producer’s interest by minimizing work-in-process inventory, inventory holding cost, and energy consumption as well as maximizing machine utilization, while the second favors the customers’ interest by maximizing customers’ service level and delivery speed. In particular, it disregards the group technology assumptions (GTAs) by allowing for the possibility of splitting pre-determined groups of jobs into inconsistent batches in order to improve the operational efficiency. A comparison between the group scheduling and batch scheduling approaches reveals the outstanding performance of the batch scheduling approach. As a result, contrary to the GTAs, jobs belonging to a group might be processed on more than one machine as batches, but not all machines may be capable of processing all jobs. A sequence- and machine-dependent setup time is required between each of two consecutively scheduled batches belonging to different groups. Based on manufacturing company policy, the desired lower bounds on batch sizes are considered for the number of jobs assigned to batches. Although, the direction in which all jobs move through production line is the same, some jobs may skip some stages. Furthermore, to reflect real industry requirements, the job release times and the machine availability times are considered to be dynamic, which means not all machines and jobs are available at the beginning of the planning horizon.The problem is formulated with the help of four mixed-integer linear programming (MILP) models. Two out of four MILP models are formulated as two integrated phases, i.e., batching and scheduling phases, with respect to the precedence constraints between each pair of jobs batches and or the position concept within batches. The optimal combination between batch compositions of groups are determined in the batching phase, while the optimal assignment and sequence of batches on machines and sequence of jobs within batches are determined in the scheduling phase, with respect to a set of operational constraints. A batch composition of a group corresponding to a particular stage, determined in the batching phase of the MILP model, represents the number of batches assigned to the group as well as the number and type of jobs belonging to each batch of that group. Since the first and second MILP models lead to unmanageable solution space, the relaxed MILP model, which allocates one and only one job to each batch of each group in each stage, can be developed to focus on the non-dominated solution space. The optimal solutions of MILP models and relaxed MILP model are equal, if and only if the optimal solution of the relaxed MILP model does not violate the desired lower bounds on batch sizes. Since the relaxed MILP model cannot guarantee the optimal solution of the MILP models, a third MILP model is developed by integrating batching and scheduling phases. This MILP model eliminates an exhaustive combination enumeration between batch compositions of all groups in all stages. Although the third MILP model converges to the optimal solution slower than the relaxed MILP model, it guarantees finding the optimal solution of the first and second MILP models. A comparison between four MILP models shows the superior performance of the third MILP model. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases from small to medium to large, even by the relaxed MILP model or the fourth MILP model. Therefore, several meta-heuristic algorithms based upon basic local search, basic population-based search, and hybridization of local search and population-based searches are developed, which move back and forth between batching and scheduling phases. Tabu Search (TS) is implemented as a basic local search algorithm, while Tabu Search Path-Relinking (TS PR) is implemented as a local search algorithm enhanced with a population-based structure. TS is incorporated into the framework of path-relinking to exploit the information on good solutions. The TS PR algorithm comprises several distinguishing features including relinking procedures to effectively explore trajectories connecting elite solutions and the methods used to choose the reference solution. Particle Swarm Optimization (PSO) is implemented as a basic population-based algorithm, while Particle Swarm Optimization enhanced with a local search algorithm (PSO LSA) is developed to realize the benefits of batching and, consequently, enhance the quality of solutions.Since there is interdependency between positions of a job in different stages of a hybrid flow shop in batch scheduling, a meta-heuristic algorithm is not capable of capturing these interdependencies and, subsequently, its efficacy can be diminished. In order to capture this interdependency, the non-, partial- complete-, and stage-based interdependency strategy are developed. In the stage-based-interdependency strategy, a complete sequence related to all of the stages is gradually determined, stage by stage. An initial solution finding mechanism is developed to trigger the search into the solution space and generate an initial population. The performances of these algorithms are compared to each other in order to identify which algorithm(s) outperforms the others. Nevertheless, the performances of the best algorithm(s) are evaluated with respect to a tight lower bound obtained from a branch-and-price (B&P) algorithm. The B&P algorithm uses Dantzig-Wolfe decomposition (DWD) to divide the original problem into a master problem and several sub-problems (SPs) corresponding to each stage. The original problem is decomposed into the SPs by three DWDs corresponding to the three MILP models. Although, by applying DWD technique in the first and second MILP models, an exhaustive combination enumeration between batch compositions of all groups in all stages is excluded and, as a result, the SPs are easier to solve than the original problem, they are still strongly NP-hard because of an enormous number of combinations between batch compositions of all groups in each stage. However, the DWD technique corresponding to the relaxed MILP model not only drastically reduces the number of variables and constraints in the SPs, but also eliminates the batching phase of the first and second MILP models. Decomposing the original problem based on the relaxed MILP model and implementing the B&P algorithm cannot guarantee optimal solutions or tight lower bounds of problems unless the number of violations in the desired lower bounds on batch sizes is not significant. Therefore, the third MILP model is decomposed by DWD so that the B&P algorithm is capable of finding tight lower bounds even for large-size instances of the problem. A comparison between the lower bounds obtained from the B&P algorithm and CPLEX reveals the impressive performance of the B&P algorithm, particularly for large-size problems. The evaluation of the best algorithms based upon these tight lower bounds developed by the B&P algorithm, uncovers the outstanding performance of hybrid algorithms compared to the results obtained from CPLEX.Keywords: Dantzig-Wolfe Decomposition, Mixed-Integer Linear Programming Model, Branch-and-Price Optimization Algorithm, Sequence- and Machine-Dependent Setup Time, Column Generation, Group Scheduling, Particle Swarm Optimization, Batching and Scheduling, Hybrid Flow Shop, Tabu Search, Desired Lower Bounds on Batch Sizes, Bi-Criteria Objective, Path-Relinkin