1,182 research outputs found

    A decomposition heuristics based on multi-bottleneck machines for large-scale job shop scheduling problems

    Get PDF
    Purpose: A decomposition heuristics based on multi-bottleneck machines for large-scale job shop scheduling problems (JSP) is proposed. Design/methodology/approach: In the algorithm, a number of sub-problems are constructed by iteratively decomposing the large-scale JSP according to the process route of each job. And then the solution of the large-scale JSP can be obtained by iteratively solving the sub-problems. In order to improve the sub-problems' solving efficiency and the solution quality, a detection method for multi-bottleneck machines based on critical path is proposed. Therewith the unscheduled operations can be decomposed into bottleneck operations and non-bottleneck operations. According to the principle of “Bottleneck leads the performance of the whole manufacturing system” in TOC (Theory Of Constraints), the bottleneck operations are scheduled by genetic algorithm for high solution quality, and the non-bottleneck operations are scheduled by dispatching rules for the improvement of the solving efficiency. Findings: In the process of the subproblems' construction, partial operations in the previous scheduled sub-problem are divided into the successive sub-problem for re-optimization. This strategy can improve the solution quality of the algorithm. In the process of solving the sub problems, the strategy that evaluating the chromosome's fitness by predicting the global scheduling objective value can improve the solution quality. Research limitations/implications: In this research, there are some assumptions which reduce the complexity of the large-scale scheduling problem. They are as follows: The processing route of each job is predetermined, and the processing time of each operation is fixed. There is no machine breakdown, and no preemption of the operations is allowed. The assumptions should be considered if the algorithm is used in the actual job shop. Originality/value: The research provides an efficient scheduling method for the large-scale job shops, and will be helpful for the discrete manufacturing industry for improving the production efficiency and effectiveness.Peer Reviewe

    A decomposition heuristics based on multi-bottleneck machines for large-scale job shop scheduling problems

    Get PDF
    Purpose: A decomposition heuristics based on multi-bottleneck machines for large-scale job shop scheduling problems (JSP) is proposed. Design/methodology/approach: In the algorithm, a number of sub-problems are constructed by iteratively decomposing the large-scale JSP according to the process route of each job. And then the solution of the large-scale JSP can be obtained by iteratively solving the sub-problems. In order to improve the sub-problems' solving efficiency and the solution quality, a detection method for multi-bottleneck machines based on critical path is proposed. Therewith the unscheduled operations can be decomposed into bottleneck operations and non-bottleneck operations. According to the principle of “Bottleneck leads the performance of the whole manufacturing system” in TOC (Theory Of Constraints), the bottleneck operations are scheduled by genetic algorithm for high solution quality, and the non-bottleneck operations are scheduled by dispatching rules for the improvement of the solving efficiency. Findings: In the process of the subproblems' construction, partial operations in the previous scheduled sub-problem are divided into the successive sub-problem for re-optimization. This strategy can improve the solution quality of the algorithm. In the process of solving the sub problems, the strategy that evaluating the chromosome's fitness by predicting the global scheduling objective value can improve the solution quality. Research limitations/implications: In this research, there are some assumptions which reduce the complexity of the large-scale scheduling problem. They are as follows: The processing route of each job is predetermined, and the processing time of each operation is fixed. There is no machine breakdown, and no preemption of the operations is allowed. The assumptions should be considered if the algorithm is used in the actual job shop. Originality/value: The research provides an efficient scheduling method for the large-scale job shops, and will be helpful for the discrete manufacturing industry for improving the production efficiency and effectiveness.Peer Reviewe

    Intelligent shop scheduling for semiconductor manufacturing

    Get PDF
    Semiconductor market sales have expanded massively to more than 200 billion dollars annually accompanied by increased pressure on the manufacturers to provide higher quality products at lower cost to remain competitive. Scheduling of semiconductor manufacturing is one of the keys to increasing productivity, however the complexity of manufacturing high capacity semiconductor devices and the cost considerations mean that it is impossible to experiment within the facility. There is an immense need for effective decision support models, characterizing and analyzing the manufacturing process, allowing the effect of changes in the production environment to be predicted in order to increase utilization and enhance system performance. Although many simulation models have been developed within semiconductor manufacturing very little research on the simulation of the photolithography process has been reported even though semiconductor manufacturers have recognized that the scheduling of photolithography is one of the most important and challenging tasks due to complex nature of the process. Traditional scheduling techniques and existing approaches show some benefits for solving small and medium sized, straightforward scheduling problems. However, they have had limited success in solving complex scheduling problems with stochastic elements in an economic timeframe. This thesis presents a new methodology combining advanced solution approaches such as simulation, artificial intelligence, system modeling and Taguchi methods, to schedule a photolithography toolset. A new structured approach was developed to effectively support building the simulation models. A single tool and complete toolset model were developed using this approach and shown to have less than 4% deviation from actual production values. The use of an intelligent scheduling agent for the toolset model shows an average of 15% improvement in simulated throughput time and is currently in use for scheduling the photolithography toolset in a manufacturing plant

    Improving project management planning and control in service operations environment.

    Get PDF
    Projects have evidently become the core activity in most companies and organisations where they are investing significant amount of resources in different types of projects as building new services, process improvement, etc. This research has focused on service sector in attempt to improve project management planning and control activities. The research is concerned with improving the planning and control of software development projects. Existing software development models are analysed and their best practices identified and these have been used to build the proposed model in this research. The research extended the existing planning and control approaches by considering uncertainty in customer requirements, resource flexibility and risks level variability. In considering these issues, the research has adopted lean principles for planning and control software development projects. A novel approach introduced within this research through the integration of simulation modelling techniques with Taguchi analysis to investigate ‗what if‘ project scenarios. Such scenarios reflect the different combinations of the factors affecting project completion time and deliverables. In addition, the research has adopted the concept of Quality Function Deployment (QFD) to develop an automated Operations Project Management Deployment (OPMD) model. The model acts as an iterative manner uses ‗what if‘ scenario performance outputs to identify constraints that may affect the completion of a certain task or phase. Any changes made during the project phases will then automatically update the performance metrics for each software development phases. In addition, optimisation routines have been developed that can be used to provide management response and to react to the different levels of uncertainty. Therefore, this research has looked at providing a comprehensive and visual overview of important project tasks i.e. progress, scheduled work, different resources, deliverables and completion that will make it easier for project members to communicate with each other to reach consensus on goals, status and required changes. Risk is important aspect that has been included in the model as well to avoid failure. The research emphasised on customer involvement, top management involvement as well as team members to be among the operational factors that escalate variability levels 3 and effect project completion time and deliverables. Therefore, commitment from everyone can improve chances of success. Although the role of different project management techniques to implement projects successfully has been widely established in areas such as the planning and control of time, cost and quality; still, the distinction between the project and project management is less than precise and a little was done in investigating different levels of uncertainty and risk levels that may occur during different project phase.United Arab Emirates Governmen

    Symmetry-Adapted Machine Learning for Information Security

    Get PDF
    Symmetry-adapted machine learning has shown encouraging ability to mitigate the security risks in information and communication technology (ICT) systems. It is a subset of artificial intelligence (AI) that relies on the principles of processing future events by learning past events or historical data. The autonomous nature of symmetry-adapted machine learning supports effective data processing and analysis for security detection in ICT systems without the interference of human authorities. Many industries are developing machine-learning-adapted solutions to support security for smart hardware, distributed computing, and the cloud. In our Special Issue book, we focus on the deployment of symmetry-adapted machine learning for information security in various application areas. This security approach can support effective methods to handle the dynamic nature of security attacks by extraction and analysis of data to identify hidden patterns of data. The main topics of this Issue include malware classification, an intrusion detection system, image watermarking, color image watermarking, battlefield target aggregation behavior recognition model, IP camera, Internet of Things (IoT) security, service function chain, indoor positioning system, and crypto-analysis

    The Use of Persistent Explorer Artificial Ants to Solve the Car Sequencing Problem

    Get PDF
    Ant Colony Optimisation is a widely researched meta-heuristic which uses the behaviour and pheromone laying activities of foraging ants to find paths through graphs. Since the early 1990’s this approach has been applied to problems such as the Travelling Salesman Problem, Quadratic Assignment Problem and Car Sequencing Problem to name a few. The ACO is not without its problems it tends to find good local optima and not good global optima. To solve this problem modifications have been made to the original ACO such as the Max Min ant system. Other solutions involve combining it with Evolutionary Algorithms to improve results. These improvements focused on the pheromone structures. Inspired by other swarm intelligence algorithms this work attempts to develop a new type of ant to explore different problem paths and thus improve the algorithm. The exploring ant would persist throughout the running time of the algorithm and explore unused paths. The Car Sequencing problem was chosen as a method to test the Exploring Ants. An existing algorithm was modified to implement the explorers. The results show that for the car sequencing problem the exploring ants did not have any positive impact, as the paths they chose were always sub-optimal

    Design of tool management systems for flexible manufacturing systems

    Get PDF
    The objective of this thesis is to study the design and analysis of tool management system in the automated manufacturing systems. The thesis is focused on two main areas, namely design and experiment. In the first part of the thesis, the design facility created has been reported. The model has been designed using a hybrid approach in which the power of both algorithmic and knowledge based approaches is utilised. Model permits detail, more accurate and complete solutions for the management of tools in a generic manufacturing system. In the second part of the thesis, to add more understanding to the tool management problems, the interactions of the major tool management design parameters have been investigated using a well known design technique, the Taguchi method. For this purpose, a large number of design experiments have been configured where some have been suggested by the Taguchi method and some have been created by the author to add more confidence, using a large body of real industrial data. The experiments results give deeper understanding of TMS problems and allow design guide-lines to be drawn for the designers. The design approach and the experiments have been proven to be an accurate and valid tool for the design of tool management systems for automated manufacturing systems. This is indicated in the conclusion of the thesis
    • 

    corecore