37 research outputs found
Serial-batch scheduling – the special case of laser-cutting machines
The dissertation deals with a problem in the field of short-term production planning, namely the scheduling of laser-cutting machines. The object of decision is the grouping of production orders (batching) and the sequencing of these order groups on one or more machines (scheduling). This problem is also known in the literature as "batch scheduling problem" and belongs to the class of combinatorial optimization problems due to the interdependencies between the batching and the scheduling decisions. The concepts and methods used are mainly from production planning, operations research and machine learning
A common framework and taxonomy for multicriteria scheduling problems with Interfering and competing Jobs: Multi-agent scheduling problems
Most classical scheduling research assumes that the objectives sought are common to all jobs to be
scheduled. However, many real-life applications can be modeled by considering different sets of jobs,
each one with its own objective(s), and an increasing number of papers addressing these problems has
appeared over the last few years. Since so far the area lacks a uni ed view, the studied problems
have received different names (such as interfering jobs, multi-agent scheduling, mixed-criteria, etc), some
authors do not seem to be aware of important contributions in related problems, and solution procedures
are often developed without taking into account existing ones. Therefore, the topic is in need of a common
framework that allows for a systematic recollection of existing contributions, as well as a clear de nition
of the main research avenues. In this paper we review multicriteria scheduling problems involving two or
more sets of jobs and propose an uni ed framework providing a common de nition, name and notation
for these problems. Moreover, we systematically review and classify the existing contributions in terms
of the complexity of the problems and the proposed solution procedures, discuss the main advances, and
point out future research lines in the topic
A survey of scheduling problems with setup times or costs
Author name used in this publication: C. T. NgAuthor name used in this publication: T. C. E. Cheng2007-2008 > Academic research: refereed > Publication in refereed journalAccepted ManuscriptPublishe
Planning and Scheduling Optimization
Although planning and scheduling optimization have been explored in the literature for many years now, it still remains a hot topic in the current scientific research. The changing market trends, globalization, technical and technological progress, and sustainability considerations make it necessary to deal with new optimization challenges in modern manufacturing, engineering, and healthcare systems. This book provides an overview of the recent advances in different areas connected with operations research models and other applications of intelligent computing techniques used for planning and scheduling optimization. The wide range of theoretical and practical research findings reported in this book confirms that the planning and scheduling problem is a complex issue that is present in different industrial sectors and organizations and opens promising and dynamic perspectives of research and development
Recommended from our members
Bi-Criteria Batching and Scheduling in Hybrid Flow Shops
In this research, a bi-criteria batching and scheduling problem is investigated in hybrid flow shop environments, where unrelated-parallel machines are run simultaneously with different capacities and eligibilities in processing, in some stages. The objective is to simultaneously minimize a linear combination of the total weighted completion time and total weighted tardiness. The first favors the producer’s interest by minimizing work-in-process inventory, inventory holding cost, and energy consumption as well as maximizing machine utilization, while the second favors the customers’ interest by maximizing customers’ service level and delivery speed. In particular, it disregards the group technology assumptions (GTAs) by allowing for the possibility of splitting pre-determined groups of jobs into inconsistent batches in order to improve the operational efficiency. A comparison between the group scheduling and batch scheduling approaches reveals the outstanding performance of the batch scheduling approach. As a result, contrary to the GTAs, jobs belonging to a group might be processed on more than one machine as batches, but not all machines may be capable of processing all jobs. A sequence- and machine-dependent setup time is required between each of two consecutively scheduled batches belonging to different groups. Based on manufacturing company policy, the desired lower bounds on batch sizes are considered for the number of jobs assigned to batches. Although, the direction in which all jobs move through production line is the same, some jobs may skip some stages. Furthermore, to reflect real industry requirements, the job release times and the machine availability times are considered to be dynamic, which means not all machines and jobs are available at the beginning of the planning horizon.The problem is formulated with the help of four mixed-integer linear programming (MILP) models. Two out of four MILP models are formulated as two integrated phases, i.e., batching and scheduling phases, with respect to the precedence constraints between each pair of jobs batches and or the position concept within batches. The optimal combination between batch compositions of groups are determined in the batching phase, while the optimal assignment and sequence of batches on machines and sequence of jobs within batches are determined in the scheduling phase, with respect to a set of operational constraints. A batch composition of a group corresponding to a particular stage, determined in the batching phase of the MILP model, represents the number of batches assigned to the group as well as the number and type of jobs belonging to each batch of that group. Since the first and second MILP models lead to unmanageable solution space, the relaxed MILP model, which allocates one and only one job to each batch of each group in each stage, can be developed to focus on the non-dominated solution space. The optimal solutions of MILP models and relaxed MILP model are equal, if and only if the optimal solution of the relaxed MILP model does not violate the desired lower bounds on batch sizes. Since the relaxed MILP model cannot guarantee the optimal solution of the MILP models, a third MILP model is developed by integrating batching and scheduling phases. This MILP model eliminates an exhaustive combination enumeration between batch compositions of all groups in all stages. Although the third MILP model converges to the optimal solution slower than the relaxed MILP model, it guarantees finding the optimal solution of the first and second MILP models. A comparison between four MILP models shows the superior performance of the third MILP model. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases from small to medium to large, even by the relaxed MILP model or the fourth MILP model. Therefore, several meta-heuristic algorithms based upon basic local search, basic population-based search, and hybridization of local search and population-based searches are developed, which move back and forth between batching and scheduling phases. Tabu Search (TS) is implemented as a basic local search algorithm, while Tabu Search Path-Relinking (TS PR) is implemented as a local search algorithm enhanced with a population-based structure. TS is incorporated into the framework of path-relinking to exploit the information on good solutions. The TS PR algorithm comprises several distinguishing features including relinking procedures to effectively explore trajectories connecting elite solutions and the methods used to choose the reference solution. Particle Swarm Optimization (PSO) is implemented as a basic population-based algorithm, while Particle Swarm Optimization enhanced with a local search algorithm (PSO LSA) is developed to realize the benefits of batching and, consequently, enhance the quality of solutions.Since there is interdependency between positions of a job in different stages of a hybrid flow shop in batch scheduling, a meta-heuristic algorithm is not capable of capturing these interdependencies and, subsequently, its efficacy can be diminished. In order to capture this interdependency, the non-, partial- complete-, and stage-based interdependency strategy are developed. In the stage-based-interdependency strategy, a complete sequence related to all of the stages is gradually determined, stage by stage. An initial solution finding mechanism is developed to trigger the search into the solution space and generate an initial population. The performances of these algorithms are compared to each other in order to identify which algorithm(s) outperforms the others. Nevertheless, the performances of the best algorithm(s) are evaluated with respect to a tight lower bound obtained from a branch-and-price (B&P) algorithm. The B&P algorithm uses Dantzig-Wolfe decomposition (DWD) to divide the original problem into a master problem and several sub-problems (SPs) corresponding to each stage. The original problem is decomposed into the SPs by three DWDs corresponding to the three MILP models. Although, by applying DWD technique in the first and second MILP models, an exhaustive combination enumeration between batch compositions of all groups in all stages is excluded and, as a result, the SPs are easier to solve than the original problem, they are still strongly NP-hard because of an enormous number of combinations between batch compositions of all groups in each stage. However, the DWD technique corresponding to the relaxed MILP model not only drastically reduces the number of variables and constraints in the SPs, but also eliminates the batching phase of the first and second MILP models. Decomposing the original problem based on the relaxed MILP model and implementing the B&P algorithm cannot guarantee optimal solutions or tight lower bounds of problems unless the number of violations in the desired lower bounds on batch sizes is not significant. Therefore, the third MILP model is decomposed by DWD so that the B&P algorithm is capable of finding tight lower bounds even for large-size instances of the problem. A comparison between the lower bounds obtained from the B&P algorithm and CPLEX reveals the impressive performance of the B&P algorithm, particularly for large-size problems. The evaluation of the best algorithms based upon these tight lower bounds developed by the B&P algorithm, uncovers the outstanding performance of hybrid algorithms compared to the results obtained from CPLEX.Keywords: Bi-Criteria Objective, Column Generation, Batch Scheduling, Tabu Search, Batching and Scheduling, Desired Lower Bounds on Batch Sizes, Path-Relinking, Branch-and-Price Optimization Algorithm, Particle Swarm Optimization, Group Scheduling, Hybrid Flow Shop, Dantzig-Wolfe Decomposition, Mixed-Integer Linear Programming Model, Sequence- and Machine-Dependent Setup Tim
Modelling and Optimizing Supply Chain Integrated Production Scheduling Problems
Globalization and advanced information technologies (e.g., Internet of Things) have considerably impacted supply chains (SCs) by persistently forcing original equipment manufacturers (OEMs) to switch production strategies from make-to-stock (MTS) to make-to-order (MTO) to survive in competition. Generally, an OEM follows the MTS strategy for products with steady demand. In contrast, the MTO strategy exists under a pull system with irregular demand in which the received customer orders are scheduled and launched into production. In comparison to MTS, MTO has the primary challenges of ensuring timely delivery at the lowest possible cost, satisfying the demands of high customization and guaranteeing the accessibility of raw materials throughout the production process. These challenges are increasing substantially since industrial productions are becoming more flexible, diversified, and customized. Besides, independently making the production scheduling decisions from other stages of these SCs often find sub-optimal results, creating substantial challenges to fulfilling demands timely and cost-effectively. Since adequately managing these challenges asynchronously are difficult, constructing optimization models by integrating SC decisions, such as customer requirements, supply portfolio (supplier selection and order allocation), delivery batching decisions, and inventory portfolio (inventory replenishment, consumption, and availability), with shop floor scheduling under a deterministic and dynamic environment is essential to fulfilling customer expectations at the least possible cost. These optimization models are computationally intractable. Consequently, designing algorithms to schedule or reschedule promptly is also highly challenging for these time-sensitive, operationally integrated optimization models. Thus, this thesis focuses on modelling and optimizing SC-integrated production scheduling problems, named SC scheduling problems (SCSPs).
The objective of optimizing job shop scheduling problems (JSSPs) is to ensure that the requisite resources are accessible when required and that their utilization is maximally efficient. Although numerous algorithms have been devised, they can sometimes become computationally exorbitant and yield sub-optimal outcomes, rendering production systems inefficient. These could be due to a variety of causes, such as an imbalance in population quality over generations, recurrent generation and evaluation of identical schedules, and permitting an under-performing method to conduct the evolutionary process. Consequently, this study designs two methods, a sequential approach (Chapter 2) and a multi-method approach (Chapter 3), to address the aforementioned issues and to acquire competitive results in finding optimal or near-optimal solutions for JSSPs in a single objective setting. The devised algorithms for JSSPs optimize workflows for each job by accurate mapping between/among related resources, generating more optimal results than existing algorithms.
Production scheduling can not be accomplished precisely without considering supply and delivery decisions and customer requirements simultaneously. Thus, a few recent studies have operationally integrated SCs to accurately predict process insights for executing, monitoring, and controlling the planned production. However, these studies are limited to simple shop-floor configurations and can provide the least flexibility to address the MTO-based SC challenges. Thus, this study formulates a bi-objective optimization model that integrates the supply portfolio into a flexible job shop scheduling environment with a customer-imposed delivery window to cost-effectively meet customized and on-time delivery requirements (Chapter 4). Compared to the job shop that is limited to sequence flexibility only, the flexible job shop has been deemed advantageous due to its capacity to provide increased scheduling flexibility (both process and sequence flexibility). To optimize the model, the performance of the multi-objective particle swarm optimization algorithm has been enhanced, with the results providing decision-makers with an increased degree of flexibility, offering a larger number of Pareto solutions, more varied and consistent frontiers, and a reasonable time for MTO-based SCs.
Environmental sustainability is spotlighted for increasing environmental awareness and follow-up regulations. Consequently, the related factors strongly regulate the supply portfolio for sustainable development, which remained unexplored in the SCSP as those criteria are primarily qualitative (e.g., green production, green product design, corporate social responsibility, and waste disposal system). These absences may lead to an unacceptable supply portfolio. Thus, this study overcomes the problem by integrating VIKORSORT into the proposed solution methodology of the extended SCSP. In addition, forming delivery batches of heterogeneous customer orders is challenging, as one order can lead to another being delayed. Therefore, the previous optimization model is extended by integrating supply, manufacturing, and delivery batching decisions and concurrently optimizing them in response to heterogeneous customer requirements with time window constraints, considering both economic and environmental sustainability for the supply portfolio (Chapter 5). Since the proposed optimization model is an extension of the flexible job shop, it can be classified as a non-deterministic polynomial-time (NP)-hard problem, which cannot be solved by conventional optimization techniques, particularly in the case of larger instances. Therefore, a reinforcement learning-based hyper-heuristic (HH) has been designed, where four solution-updating heuristics are intelligently guided to deliver the best possible results compared to existing algorithms. The optimization model furnishes a set of comprehensive schedules that integrate the supply portfolio, production portfolio (work-center/machine assignment and customer orders sequencing), and batching decisions. This provides numerous meaningful managerial insights and operational flexibility prior to the execution phase.
Recently, SCs have been experiencing unprecedented and massive disruptions caused by an abrupt outbreak, resulting in difficulties for OEMs to recover from disruptive demand-supply equilibrium. Hence, this study proposes a multi-portfolio (supply, production, and inventory portfolios) approach for a proactive-reactive scheme, which concerns the SCSP with complex multi-level products, simultaneously including unpredictably dynamic supply, demand, and shop floor disruptions (Chapter 6). This study considers fabrication and assembly in a multi-level product structure. To effectively address this time-sensitive model based on real-time data, a Q-learning-based multi-operator differential evolution algorithm in a HH has been designed to address disruptive events and generate a timely rescheduling plan. The numerical results and analyses demonstrate the proposed model's capability to effectively address single and multiple disruptions, thus providing significant managerial insights and ensuring SC resilience
Recommended from our members
Bi-Criteria Batching and Scheduling in Hybrid Flow Shops
In this research, a bi-criteria batching and scheduling problem is investigated in hybrid flow shop environments, where unrelated-parallel machines are run simultaneously with different capacities and eligibilities in processing, in some stages. The objective is to simultaneously minimize a linear combination of the total weighted completion time and total weighted tardiness. The first favors the producer’s interest by minimizing work-in-process inventory, inventory holding cost, and energy consumption as well as maximizing machine utilization, while the second favors the customers’ interest by maximizing customers’ service level and delivery speed. In particular, it disregards the group technology assumptions (GTAs) by allowing for the possibility of splitting pre-determined groups of jobs into inconsistent batches in order to improve the operational efficiency. A comparison between the group scheduling and batch scheduling approaches reveals the outstanding performance of the batch scheduling approach. As a result, contrary to the GTAs, jobs belonging to a group might be processed on more than one machine as batches, but not all machines may be capable of processing all jobs. A sequence- and machine-dependent setup time is required between each of two consecutively scheduled batches belonging to different groups. Based on manufacturing company policy, the desired lower bounds on batch sizes are considered for the number of jobs assigned to batches. Although, the direction in which all jobs move through production line is the same, some jobs may skip some stages. Furthermore, to reflect real industry requirements, the job release times and the machine availability times are considered to be dynamic, which means not all machines and jobs are available at the beginning of the planning horizon.The problem is formulated with the help of four mixed-integer linear programming (MILP) models. Two out of four MILP models are formulated as two integrated phases, i.e., batching and scheduling phases, with respect to the precedence constraints between each pair of jobs batches and or the position concept within batches. The optimal combination between batch compositions of groups are determined in the batching phase, while the optimal assignment and sequence of batches on machines and sequence of jobs within batches are determined in the scheduling phase, with respect to a set of operational constraints. A batch composition of a group corresponding to a particular stage, determined in the batching phase of the MILP model, represents the number of batches assigned to the group as well as the number and type of jobs belonging to each batch of that group. Since the first and second MILP models lead to unmanageable solution space, the relaxed MILP model, which allocates one and only one job to each batch of each group in each stage, can be developed to focus on the non-dominated solution space. The optimal solutions of MILP models and relaxed MILP model are equal, if and only if the optimal solution of the relaxed MILP model does not violate the desired lower bounds on batch sizes. Since the relaxed MILP model cannot guarantee the optimal solution of the MILP models, a third MILP model is developed by integrating batching and scheduling phases. This MILP model eliminates an exhaustive combination enumeration between batch compositions of all groups in all stages. Although the third MILP model converges to the optimal solution slower than the relaxed MILP model, it guarantees finding the optimal solution of the first and second MILP models. A comparison between four MILP models shows the superior performance of the third MILP model. However, since the problem is strongly NP-hard, it is not possible to find its optimal solution within a reasonable time as the problem size increases from small to medium to large, even by the relaxed MILP model or the fourth MILP model. Therefore, several meta-heuristic algorithms based upon basic local search, basic population-based search, and hybridization of local search and population-based searches are developed, which move back and forth between batching and scheduling phases. Tabu Search (TS) is implemented as a basic local search algorithm, while Tabu Search Path-Relinking (TS PR) is implemented as a local search algorithm enhanced with a population-based structure. TS is incorporated into the framework of path-relinking to exploit the information on good solutions. The TS PR algorithm comprises several distinguishing features including relinking procedures to effectively explore trajectories connecting elite solutions and the methods used to choose the reference solution. Particle Swarm Optimization (PSO) is implemented as a basic population-based algorithm, while Particle Swarm Optimization enhanced with a local search algorithm (PSO LSA) is developed to realize the benefits of batching and, consequently, enhance the quality of solutions.Since there is interdependency between positions of a job in different stages of a hybrid flow shop in batch scheduling, a meta-heuristic algorithm is not capable of capturing these interdependencies and, subsequently, its efficacy can be diminished. In order to capture this interdependency, the non-, partial- complete-, and stage-based interdependency strategy are developed. In the stage-based-interdependency strategy, a complete sequence related to all of the stages is gradually determined, stage by stage. An initial solution finding mechanism is developed to trigger the search into the solution space and generate an initial population. The performances of these algorithms are compared to each other in order to identify which algorithm(s) outperforms the others. Nevertheless, the performances of the best algorithm(s) are evaluated with respect to a tight lower bound obtained from a branch-and-price (B&P) algorithm. The B&P algorithm uses Dantzig-Wolfe decomposition (DWD) to divide the original problem into a master problem and several sub-problems (SPs) corresponding to each stage. The original problem is decomposed into the SPs by three DWDs corresponding to the three MILP models. Although, by applying DWD technique in the first and second MILP models, an exhaustive combination enumeration between batch compositions of all groups in all stages is excluded and, as a result, the SPs are easier to solve than the original problem, they are still strongly NP-hard because of an enormous number of combinations between batch compositions of all groups in each stage. However, the DWD technique corresponding to the relaxed MILP model not only drastically reduces the number of variables and constraints in the SPs, but also eliminates the batching phase of the first and second MILP models. Decomposing the original problem based on the relaxed MILP model and implementing the B&P algorithm cannot guarantee optimal solutions or tight lower bounds of problems unless the number of violations in the desired lower bounds on batch sizes is not significant. Therefore, the third MILP model is decomposed by DWD so that the B&P algorithm is capable of finding tight lower bounds even for large-size instances of the problem. A comparison between the lower bounds obtained from the B&P algorithm and CPLEX reveals the impressive performance of the B&P algorithm, particularly for large-size problems. The evaluation of the best algorithms based upon these tight lower bounds developed by the B&P algorithm, uncovers the outstanding performance of hybrid algorithms compared to the results obtained from CPLEX.Keywords: Dantzig-Wolfe Decomposition, Mixed-Integer Linear Programming Model, Branch-and-Price Optimization Algorithm, Sequence- and Machine-Dependent Setup Time, Column Generation, Group Scheduling, Particle Swarm Optimization, Batching and Scheduling, Hybrid Flow Shop, Tabu Search, Desired Lower Bounds on Batch Sizes, Bi-Criteria Objective, Path-Relinkin
Advances and Novel Approaches in Discrete Optimization
Discrete optimization is an important area of Applied Mathematics with a broad spectrum of applications in many fields. This book results from a Special Issue in the journal Mathematics entitled ‘Advances and Novel Approaches in Discrete Optimization’. It contains 17 articles covering a broad spectrum of subjects which have been selected from 43 submitted papers after a thorough refereeing process. Among other topics, it includes seven articles dealing with scheduling problems, e.g., online scheduling, batching, dual and inverse scheduling problems, or uncertain scheduling problems. Other subjects are graphs and applications, evacuation planning, the max-cut problem, capacitated lot-sizing, and packing algorithms