9 research outputs found
Optimasi Penjadwalan Flow Shop Menggunakan Algoritma Hybrid Differential Evolution
Penjadwalan produksi merupakan bagian integral di dalam sistem manufaktur. Artikel ini menyelesaikan permasalahan penjadwalan flow shop dengan fungsi obyektif total flow time. Dalam penjadwalan, total flow time menghasilkan konsumsi yang stabil terhadap sumber daya, perputaran job yang cepat serta meminimalkan work in process inventory. Permasalahan penjadwalan flow shop tergolong pada permasalahan optimasi kombinatorial yang merupakan NP-hard. Saat ini, penggunaan algoritma metaheuristik banyak digunakan untuk memecahkan kasus optimasi kombinatorial, termasuk penjadwalan flow shop. Salah satu yang memiliki performa yang baik adalah Algoritma Differential Evolution. Untuk meningkatkan kualitas solusinya, Algoritma Differential Evolution akan ditambahkan dengan prosedur local search yang dinamakan Hybrid Differential Evolution. Untuk mengetahui performa dari algoritma tersebut, dilakukan pengujian menggunakan data penjadwalan flow shop yang ada pada OR-Library. Performa Hybrid Differential Evolution akan dibandingkan dengan algoritma lain. Hasil pengujian menunjukkan bahwa Hybrid Differential Evolution memberikan performa yang lebih baik dibandingkan dengan algoritma lain
a hybrid metaheuristic approach for minimizing the total flow time in a flow shop sequence dependent group scheduling problem
Production processes in Cellular Manufacturing Systems (CMS) often involve groups of parts sharing the same technological requirements in terms of tooling and setup. The issue of scheduling such parts through a flow-shop production layout is known as the Flow-Shop Group Scheduling (FSGS) problem or, whether setup times are sequence-dependent, the Flow-Shop Sequence-Dependent Group Scheduling (FSDGS) problem. This paper addresses the FSDGS issue, proposing a hybrid metaheuristic procedure integrating features from Genetic Algorithms (GAs) and Biased Random Sampling (BRS) search techniques with the aim of minimizing the total flow time, i.e., the sum of completion times of all jobs. A well-known benchmark of test cases, entailing problems with two, three, and six machines, is employed for both tuning the relevant parameters of the developed procedure and assessing its performances against two metaheuristic algorithms recently presented by literature. The obtained results and a properly arranged ANOVA analysis highlight the superiority of the proposed approach in tackling the scheduling problem under investigation
Recommended from our members
Sequence-dependent group scheduling problem on unrelated-parallel machines
In this research we address a sequence-dependent group scheduling problem on a set of unrelated-parallel machines where the run time of each job differs on different machines. To benefit both producer and customers we attempt to minimize a linear combination of total weighted completion time and total weighted tardiness. Since the problem is shown to be NP-hard, meta-heuristic algorithms based on tabu search are developed to find the optimal/near optimal solution. For some small size yet complex problems, the results from these algorithms are compared to the optimal solutions found by CPLEX. The result obtained in all of these problems is that the tabu search algorithms could find solutions at least as good as CPLEX but in drastically shorter computational time, thus signifying the high degree of efficiency and efficacy attained by the former.Keywords: Bi-criteria, Group scheduling, Sequence-dependent setup time, Mixed-integer linear programming, Unrelated-parallel machines, Tabu searc
Problem specific heuristics for group scheduling problems in cellular manufacturing
The group scheduling problem commonly arises in cellular manufacturing systems, where parts are grouped into part families. It is characterized by a sequencing task on two levels: on the one hand, a sequence of jobs within each part family has to be identified while, on the other hand, a family sequence has to be determined. In order to solve this NP-hard problem usually heuristic solution approaches are used. In this thesis different aspects of group scheduling are discussed and problem specific heuristics are developed to solve group scheduling problems efficiently. Thereby, particularly characteristic properties of flowshop group scheduling problems, such as the structure of a group schedule or missing operations, are identified and exploited. In a simulation study for job shop manufacturing cells several novel dispatching rules are analyzed. Furthermore, a comprehensive review of the existing group scheduling literature is presented, identifying fruitful directions for future research
Recommended from our members
Bi-criteria group scheduling with learning in hybrid flow shops
In this research, a bi-criteria group scheduling problem is investigated in hybrid flow shop (HFS) environments, where the parallel machines in each stage are unrelated, meaning not identical. The objective of the problem is to minimize a linear combination of the total weighted completion times as a means of complying with the interests of the producer, and the total weighted tardiness as a means of complying with the interests of customers. The underlying assumptions of the problem include the group technology assumptions (GTA) that require all jobs within a group to be processed successively and on the same machine. The runtime of these jobs are dynamic and progressively decrease as the worker learns how to perform similar jobs. A sequence-dependent setup time is considered for switching between different groups on the same machine. Although all jobs have to move in unidirectional paths through the HFS, some may skip some of the stages. Furthermore, in order to capture more realistic features of the scheduling problems, the jobs are assumed to be released into the system at dynamic times, and the machines, as well, are assumed to be available at dynamic times. The problem is formulated as a mixed-integer linear programming (MILP) model. The MILP model for small sizes of the problem is solved to optimality using CPLEX. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases to medium to large.
Several meta-heuristic algorithms based on tabu search (TS), simulated annealing (SA), and genetic algorithm (GA) are developed to find the optimal/near optimal solutions for this problem. Three alterations of algorithms are developed for TS and SA-based algorithms (referred to as local search algorithms) i.e. non-permutation, partial permutation and local searches with embedded progressive perturbations. Two alternatives are also considered for GA-based algorithms (referred to as population-based algorithms) i.e. simple GA and bi-level GA. The performances of these algorithms are compared to each other in order to identify which algorithm, if any, outperforms the others. Nevertheless, the performances of all algorithms are evaluated with respect to a tight lower bound (LB) obtained based on a branch-and-price (B&P) technique developed in this research.
The B&P technique uses Dantzig-Wolfe decomposition to divide the original problem into a master problem and several sub-problems. Although, the sub-problems are smaller than the original problem, they are still strongly NP-hard and cannot be optimally solved within a reasonable amount of time. However, an optimal dispatching rule is proposed that drastically reduces the number of variables and constraints in these sub-problems, and enables the B&P algorithm to find tight lower bounds even for large-size instances of the problem. A comparison between these lower bounds and the ones obtained from CPLEX reveals the impressive performance of the B&P algorithm, i.e. an average of 233% improvement for the largest size of the problems that have been tested. Evaluation of the proposed algorithms with respect to these tight lower bounds uncovers the outstanding performance of all the proposed algorithms, while identifying the bi-level GA as the best performing algorithm in dealing with the HFS scheduling problem. This algorithm reports a remarkable performance with an average deviation of only 2% from the optimal solution for small-size sample problems, and an average gap of 23% from the lower bound for the largest sizes of the tested problems. The largest problem tested in this research consists of a total of 1858 binary variables and 14654 constraints
Recommended from our members
Bi-Criteria Batching and Scheduling in Hybrid Flow Shops
In this research, a bi-criteria batching and scheduling problem is investigated in hybrid flow shop environments, where unrelated-parallel machines are run simultaneously with different capacities and eligibilities in processing, in some stages. The objective is to simultaneously minimize a linear combination of the total weighted completion time and total weighted tardiness. The first favors the producer’s interest by minimizing work-in-process inventory, inventory holding cost, and energy consumption as well as maximizing machine utilization, while the second favors the customers’ interest by maximizing customers’ service level and delivery speed. In particular, it disregards the group technology assumptions (GTAs) by allowing for the possibility of splitting pre-determined groups of jobs into inconsistent batches in order to improve the operational efficiency. A comparison between the group scheduling and batch scheduling approaches reveals the outstanding performance of the batch scheduling approach. As a result, contrary to the GTAs, jobs belonging to a group might be processed on more than one machine as batches, but not all machines may be capable of processing all jobs. A sequence- and machine-dependent setup time is required between each of two consecutively scheduled batches belonging to different groups. Based on manufacturing company policy, the desired lower bounds on batch sizes are considered for the number of jobs assigned to batches. Although, the direction in which all jobs move through production line is the same, some jobs may skip some stages. Furthermore, to reflect real industry requirements, the job release times and the machine availability times are considered to be dynamic, which means not all machines and jobs are available at the beginning of the planning horizon.The problem is formulated with the help of four mixed-integer linear programming (MILP) models. Two out of four MILP models are formulated as two integrated phases, i.e., batching and scheduling phases, with respect to the precedence constraints between each pair of jobs batches and or the position concept within batches. The optimal combination between batch compositions of groups are determined in the batching phase, while the optimal assignment and sequence of batches on machines and sequence of jobs within batches are determined in the scheduling phase, with respect to a set of operational constraints. A batch composition of a group corresponding to a particular stage, determined in the batching phase of the MILP model, represents the number of batches assigned to the group as well as the number and type of jobs belonging to each batch of that group. Since the first and second MILP models lead to unmanageable solution space, the relaxed MILP model, which allocates one and only one job to each batch of each group in each stage, can be developed to focus on the non-dominated solution space. The optimal solutions of MILP models and relaxed MILP model are equal, if and only if the optimal solution of the relaxed MILP model does not violate the desired lower bounds on batch sizes. Since the relaxed MILP model cannot guarantee the optimal solution of the MILP models, a third MILP model is developed by integrating batching and scheduling phases. This MILP model eliminates an exhaustive combination enumeration between batch compositions of all groups in all stages. Although the third MILP model converges to the optimal solution slower than the relaxed MILP model, it guarantees finding the optimal solution of the first and second MILP models. A comparison between four MILP models shows the superior performance of the third MILP model. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases from small to medium to large, even by the relaxed MILP model or the fourth MILP model. Therefore, several meta-heuristic algorithms based upon basic local search, basic population-based search, and hybridization of local search and population-based searches are developed, which move back and forth between batching and scheduling phases. Tabu Search (TS) is implemented as a basic local search algorithm, while Tabu Search Path-Relinking (TS PR) is implemented as a local search algorithm enhanced with a population-based structure. TS is incorporated into the framework of path-relinking to exploit the information on good solutions. The TS PR algorithm comprises several distinguishing features including relinking procedures to effectively explore trajectories connecting elite solutions and the methods used to choose the reference solution. Particle Swarm Optimization (PSO) is implemented as a basic population-based algorithm, while Particle Swarm Optimization enhanced with a local search algorithm (PSO LSA) is developed to realize the benefits of batching and, consequently, enhance the quality of solutions.Since there is interdependency between positions of a job in different stages of a hybrid flow shop in batch scheduling, a meta-heuristic algorithm is not capable of capturing these interdependencies and, subsequently, its efficacy can be diminished. In order to capture this interdependency, the non-, partial- complete-, and stage-based interdependency strategy are developed. In the stage-based-interdependency strategy, a complete sequence related to all of the stages is gradually determined, stage by stage. An initial solution finding mechanism is developed to trigger the search into the solution space and generate an initial population. The performances of these algorithms are compared to each other in order to identify which algorithm(s) outperforms the others. Nevertheless, the performances of the best algorithm(s) are evaluated with respect to a tight lower bound obtained from a branch-and-price (B&P) algorithm. The B&P algorithm uses Dantzig-Wolfe decomposition (DWD) to divide the original problem into a master problem and several sub-problems (SPs) corresponding to each stage. The original problem is decomposed into the SPs by three DWDs corresponding to the three MILP models. Although, by applying DWD technique in the first and second MILP models, an exhaustive combination enumeration between batch compositions of all groups in all stages is excluded and, as a result, the SPs are easier to solve than the original problem, they are still strongly NP-hard because of an enormous number of combinations between batch compositions of all groups in each stage. However, the DWD technique corresponding to the relaxed MILP model not only drastically reduces the number of variables and constraints in the SPs, but also eliminates the batching phase of the first and second MILP models. Decomposing the original problem based on the relaxed MILP model and implementing the B&P algorithm cannot guarantee optimal solutions or tight lower bounds of problems unless the number of violations in the desired lower bounds on batch sizes is not significant. Therefore, the third MILP model is decomposed by DWD so that the B&P algorithm is capable of finding tight lower bounds even for large-size instances of the problem. A comparison between the lower bounds obtained from the B&P algorithm and CPLEX reveals the impressive performance of the B&P algorithm, particularly for large-size problems. The evaluation of the best algorithms based upon these tight lower bounds developed by the B&P algorithm, uncovers the outstanding performance of hybrid algorithms compared to the results obtained from CPLEX.Keywords: Bi-Criteria Objective, Column Generation, Batch Scheduling, Tabu Search, Batching and Scheduling, Desired Lower Bounds on Batch Sizes, Path-Relinking, Branch-and-Price Optimization Algorithm, Particle Swarm Optimization, Group Scheduling, Hybrid Flow Shop, Dantzig-Wolfe Decomposition, Mixed-Integer Linear Programming Model, Sequence- and Machine-Dependent Setup Tim
Recommended from our members
Bi-Criteria Batching and Scheduling in Hybrid Flow Shops
In this research, a bi-criteria batching and scheduling problem is investigated in hybrid flow shop environments, where unrelated-parallel machines are run simultaneously with different capacities and eligibilities in processing, in some stages. The objective is to simultaneously minimize a linear combination of the total weighted completion time and total weighted tardiness. The first favors the producer’s interest by minimizing work-in-process inventory, inventory holding cost, and energy consumption as well as maximizing machine utilization, while the second favors the customers’ interest by maximizing customers’ service level and delivery speed. In particular, it disregards the group technology assumptions (GTAs) by allowing for the possibility of splitting pre-determined groups of jobs into inconsistent batches in order to improve the operational efficiency. A comparison between the group scheduling and batch scheduling approaches reveals the outstanding performance of the batch scheduling approach. As a result, contrary to the GTAs, jobs belonging to a group might be processed on more than one machine as batches, but not all machines may be capable of processing all jobs. A sequence- and machine-dependent setup time is required between each of two consecutively scheduled batches belonging to different groups. Based on manufacturing company policy, the desired lower bounds on batch sizes are considered for the number of jobs assigned to batches. Although, the direction in which all jobs move through production line is the same, some jobs may skip some stages. Furthermore, to reflect real industry requirements, the job release times and the machine availability times are considered to be dynamic, which means not all machines and jobs are available at the beginning of the planning horizon.The problem is formulated with the help of four mixed-integer linear programming (MILP) models. Two out of four MILP models are formulated as two integrated phases, i.e., batching and scheduling phases, with respect to the precedence constraints between each pair of jobs batches and or the position concept within batches. The optimal combination between batch compositions of groups are determined in the batching phase, while the optimal assignment and sequence of batches on machines and sequence of jobs within batches are determined in the scheduling phase, with respect to a set of operational constraints. A batch composition of a group corresponding to a particular stage, determined in the batching phase of the MILP model, represents the number of batches assigned to the group as well as the number and type of jobs belonging to each batch of that group. Since the first and second MILP models lead to unmanageable solution space, the relaxed MILP model, which allocates one and only one job to each batch of each group in each stage, can be developed to focus on the non-dominated solution space. The optimal solutions of MILP models and relaxed MILP model are equal, if and only if the optimal solution of the relaxed MILP model does not violate the desired lower bounds on batch sizes. Since the relaxed MILP model cannot guarantee the optimal solution of the MILP models, a third MILP model is developed by integrating batching and scheduling phases. This MILP model eliminates an exhaustive combination enumeration between batch compositions of all groups in all stages. Although the third MILP model converges to the optimal solution slower than the relaxed MILP model, it guarantees finding the optimal solution of the first and second MILP models. A comparison between four MILP models shows the superior performance of the third MILP model. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases from small to medium to large, even by the relaxed MILP model or the fourth MILP model. Therefore, several meta-heuristic algorithms based upon basic local search, basic population-based search, and hybridization of local search and population-based searches are developed, which move back and forth between batching and scheduling phases. Tabu Search (TS) is implemented as a basic local search algorithm, while Tabu Search Path-Relinking (TS PR) is implemented as a local search algorithm enhanced with a population-based structure. TS is incorporated into the framework of path-relinking to exploit the information on good solutions. The TS PR algorithm comprises several distinguishing features including relinking procedures to effectively explore trajectories connecting elite solutions and the methods used to choose the reference solution. Particle Swarm Optimization (PSO) is implemented as a basic population-based algorithm, while Particle Swarm Optimization enhanced with a local search algorithm (PSO LSA) is developed to realize the benefits of batching and, consequently, enhance the quality of solutions.Since there is interdependency between positions of a job in different stages of a hybrid flow shop in batch scheduling, a meta-heuristic algorithm is not capable of capturing these interdependencies and, subsequently, its efficacy can be diminished. In order to capture this interdependency, the non-, partial- complete-, and stage-based interdependency strategy are developed. In the stage-based-interdependency strategy, a complete sequence related to all of the stages is gradually determined, stage by stage. An initial solution finding mechanism is developed to trigger the search into the solution space and generate an initial population. The performances of these algorithms are compared to each other in order to identify which algorithm(s) outperforms the others. Nevertheless, the performances of the best algorithm(s) are evaluated with respect to a tight lower bound obtained from a branch-and-price (B&P) algorithm. The B&P algorithm uses Dantzig-Wolfe decomposition (DWD) to divide the original problem into a master problem and several sub-problems (SPs) corresponding to each stage. The original problem is decomposed into the SPs by three DWDs corresponding to the three MILP models. Although, by applying DWD technique in the first and second MILP models, an exhaustive combination enumeration between batch compositions of all groups in all stages is excluded and, as a result, the SPs are easier to solve than the original problem, they are still strongly NP-hard because of an enormous number of combinations between batch compositions of all groups in each stage. However, the DWD technique corresponding to the relaxed MILP model not only drastically reduces the number of variables and constraints in the SPs, but also eliminates the batching phase of the first and second MILP models. Decomposing the original problem based on the relaxed MILP model and implementing the B&P algorithm cannot guarantee optimal solutions or tight lower bounds of problems unless the number of violations in the desired lower bounds on batch sizes is not significant. Therefore, the third MILP model is decomposed by DWD so that the B&P algorithm is capable of finding tight lower bounds even for large-size instances of the problem. A comparison between the lower bounds obtained from the B&P algorithm and CPLEX reveals the impressive performance of the B&P algorithm, particularly for large-size problems. The evaluation of the best algorithms based upon these tight lower bounds developed by the B&P algorithm, uncovers the outstanding performance of hybrid algorithms compared to the results obtained from CPLEX.Keywords: Dantzig-Wolfe Decomposition, Mixed-Integer Linear Programming Model, Branch-and-Price Optimization Algorithm, Sequence- and Machine-Dependent Setup Time, Column Generation, Group Scheduling, Particle Swarm Optimization, Batching and Scheduling, Hybrid Flow Shop, Tabu Search, Desired Lower Bounds on Batch Sizes, Bi-Criteria Objective, Path-Relinkin