13,748 research outputs found

    A Hybrid Tabu/Scatter Search Algorithm for Simulation-Based Optimization of Multi-Objective Runway Operations Scheduling

    Get PDF
    As air traffic continues to increase, air traffic flow management is becoming more challenging to effectively and efficiently utilize airport capacity without compromising safety, environmental and economic requirements. Since runways are often the primary limiting factor in airport capacity, runway operations scheduling emerge as an important problem to be solved to alleviate flight delays and air traffic congestion while reducing unnecessary fuel consumption and negative environmental impacts. However, even a moderately sized real-life runway operations scheduling problem tends to be too complex to be solved by analytical methods, where all mathematical models for this problem belong to the complexity class of NP-Hard in a strong sense due to combinatorial nature of the problem. Therefore, it is only possible to solve practical runway operations scheduling problem by making a large number of simplifications and assumptions in a deterministic context. As a result, most analytical models proposed in the literature suffer from too much abstraction, avoid uncertainties and, in turn, have little applicability in practice. On the other hand, simulation-based methods have the capability to characterize complex and stochastic real-life runway operations in detail, and to cope with several constraints and stakeholders’ preferences, which are commonly considered as important factors in practice. This dissertation proposes a simulation-based optimization (SbO) approach for multi-objective runway operations scheduling problem. The SbO approach utilizes a discrete-event simulation model for accounting for uncertain conditions, and an optimization component for finding the best known Pareto set of solutions. This approach explicitly considers uncertainty to decrease the real operational cost of the runway operations as well as fairness among aircraft as part of the optimization process. Due to the problem’s large, complex and unstructured search space, a hybrid Tabu/Scatter Search algorithm is developed to find solutions by using an elitist strategy to preserve non-dominated solutions, a dynamic update mechanism to produce high-quality solutions and a rebuilding strategy to promote solution diversity. The proposed algorithm is applied to bi-objective (i.e., maximizing runway utilization and fairness) runway operations schedule optimization as the optimization component of the SbO framework, where the developed simulation model acts as an external function evaluator. To the best of our knowledge, this is the first SbO approach that explicitly considers uncertainties in the development of schedules for runway operations as well as considers fairness as a secondary objective. In addition, computational experiments are conducted using real-life datasets for a major US airport to demonstrate that the proposed approach is effective and computationally tractable in a practical sense. In the experimental design, statistical design of experiments method is employed to analyze the impacts of parameters on the simulation as well as on the optimization component’s performance, and to identify the appropriate parameter levels. The results show that the implementation of the proposed SbO approach provides operational benefits when compared to First-Come-First-Served (FCFS) and deterministic approaches without compromising schedule fairness. It is also shown that proposed algorithm is capable of generating a set of solutions that represent the inherent trade-offs between the objectives that are considered. The proposed decision-making algorithm might be used as part of decision support tools to aid air traffic controllers in solving the real-life runway operations scheduling problem

    Scheduling in assembly type job-shops

    Get PDF
    Assembly type job-shop scheduling is a generalization of the job-shop scheduling problem to include assembly operations. In the assembly type job-shops scheduling problem, there are n jobs which are to be processed on in workstations and each job has a due date. Each job visits one or more workstations in a predetermined route. The primary difference between this new problem and the classical job-shop problem is that two or more jobs can merge to foul\u27 a new job at a specified workstation, that is job convergence is permitted. This feature cannot be modeled by existing job-shop techniques. In this dissertation, we develop scheduling procedures for the assembly type job-shop with the objective of minimizing total weighted tardiness. Three types of workstations are modeled: single machine, parallel machine, and batch machine. We label this new scheduling procedure as SB. The SB procedure is heuristic in nature and is derived from the shifting bottleneck concept. SB decomposes the assembly type job-shop scheduling problem into several workstation scheduling sub-problems. Various types of techniques are used in developing the scheduling heuristics for these sub-problems including the greedy method, beam search, critical path analysis, local search, and dynamic programming. The performance of SB is validated on a set of test problems and compared with priority rules that are normally used in practice. The results show that SB outperforms the priority rules by an average of 19% - 36% for the test problems. SB is extended to solve scheduling problems with other objectives including minimizing the maximum completion time, minimizing weighted flow time and minimizing maximum weighted lateness. Comparisons with the test problems, indicate that SB outperforms the priority rules for these objectives as well. The SB procedure and its accompanying logic is programmed into an object oriented scheduling system labeled as LEKIN. The LEKIN program includes a standard library of scheduling rules and hence can be used as a platform for the development of new scheduling heuristics. In industrial applications LEKIN allows schedulers to obtain effective machine schedules rapidly. The results from this research allow us to increase shop utilization, improve customer satisfaction, and lower work-in-process inventory without a major capital investment

    Overlap Removal of Dimensionality Reduction Scatterplot Layouts

    Full text link
    Dimensionality Reduction (DR) scatterplot layouts have become a ubiquitous visualization tool for analyzing multidimensional data items with presence in different areas. Despite its popularity, scatterplots suffer from occlusion, especially when markers convey information, making it troublesome for users to estimate items' groups' sizes and, more importantly, potentially obfuscating critical items for the analysis under execution. Different strategies have been devised to address this issue, either producing overlap-free layouts, lacking the powerful capabilities of contemporary DR techniques in uncover interesting data patterns, or eliminating overlaps as a post-processing strategy. Despite the good results of post-processing techniques, the best methods typically expand or distort the scatterplot area, thus reducing markers' size (sometimes) to unreadable dimensions, defeating the purpose of removing overlaps. This paper presents a novel post-processing strategy to remove DR layouts' overlaps that faithfully preserves the original layout's characteristics and markers' sizes. We show that the proposed strategy surpasses the state-of-the-art in overlap removal through an extensive comparative evaluation considering multiple different metrics while it is 2 or 3 orders of magnitude faster for large datasets.Comment: 11 pages and 9 figure

    The 6th Conference of PhD Students in Computer Science

    Get PDF

    Reactive approach for automating exploration and exploitation in ant colony optimization

    Get PDF
    Ant colony optimization (ACO) algorithms can be used to solve nondeterministic polynomial hard problems. Exploration and exploitation are the main mechanisms in controlling search within the ACO. Reactive search is an alternative technique to maintain the dynamism of the mechanics. However, ACO-based reactive search technique has three (3) problems. First, the memory model to record previous search regions did not completely transfer the neighborhood structures to the next iteration which leads to arbitrary restart and premature local search. Secondly, the exploration indicator is not robust due to the difference of magnitudes in distance matrices for the current population. Thirdly, the parameter control techniques that utilize exploration indicators in their feedback process do not consider the problem of indicator robustness. A reactive ant colony optimization (RACO) algorithm has been proposed to overcome the limitations of the reactive search. RACO consists of three main components. The first component is a reactive max-min ant system algorithm for recording the neighborhood structures. The second component is a statistical machine learning mechanism named ACOustic to produce a robust exploration indicator. The third component is the ACO-based adaptive parameter selection algorithm to solve the parameterization problem which relies on quality, exploration and unified criteria in assigning rewards to promising parameters. The performance of RACO is evaluated on traveling salesman and quadratic assignment problems and compared with eight metaheuristics techniques in terms of success rate, Wilcoxon signed-rank, Chi-square and relative percentage deviation. Experimental results showed that the performance of RACO is superior than the eight (8) metaheuristics techniques which confirmed that RACO can be used as a new direction for solving optimization problems. RACO can be used in providing a dynamic exploration and exploitation mechanism, setting a parameter value which allows an efficient search, describing the amount of exploration an ACO algorithm performs and detecting stagnation situations

    Parallel optimization algorithms for high performance computing : application to thermal systems

    Get PDF
    The need of optimization is present in every field of engineering. Moreover, applications requiring a multidisciplinary approach in order to make a step forward are increasing. This leads to the need of solving complex optimization problems that exceed the capacity of human brain or intuition. A standard way of proceeding is to use evolutionary algorithms, among which genetic algorithms hold a prominent place. These are characterized by their robustness and versatility, as well as their high computational cost and low convergence speed. Many optimization packages are available under free software licenses and are representative of the current state of the art in optimization technology. However, the ability of optimization algorithms to adapt to massively parallel computers reaching satisfactory efficiency levels is still an open issue. Even packages suited for multilevel parallelism encounter difficulties when dealing with objective functions involving long and variable simulation times. This variability is common in Computational Fluid Dynamics and Heat Transfer (CFD & HT), nonlinear mechanics, etc. and is nowadays a dominant concern for large scale applications. Current research in improving the performance of evolutionary algorithms is mainly focused on developing new search algorithms. Nevertheless, there is a vast knowledge of sequential well-performing algorithmic suitable for being implemented in parallel computers. The gap to be covered is efficient parallelization. Moreover, advances in the research of both new search algorithms and efficient parallelization are additive, so that the enhancement of current state of the art optimization software can be accelerated if both fronts are tackled simultaneously. The motivation of this Doctoral Thesis is to make a step forward towards the successful integration of Optimization and High Performance Computing capabilities, which has the potential to boost technological development by providing better designs, shortening product development times and minimizing the required resources. After conducting a thorough state of the art study of the mathematical optimization techniques available to date, a generic mathematical optimization tool has been developed putting a special focus on the application of the library to the field of Computational Fluid Dynamics and Heat Transfer (CFD & HT). Then the main shortcomings of the standard parallelization strategies available for genetic algorithms and similar population-based optimization methods have been analyzed. Computational load imbalance has been identified to be the key point causing the degradation of the optimization algorithm¿s scalability (i.e. parallel efficiency) in case the average makespan of the batch of individuals is greater than the average time required by the optimizer for performing inter-processor communications. It occurs because processors are often unable to finish the evaluation of their queue of individuals simultaneously and need to be synchronized before the next batch of individuals is created. Consequently, the computational load imbalance is translated into idle time in some processors. Several load balancing algorithms have been proposed and exhaustively tested, being extendable to any other population-based optimization method that needs to synchronize all processors after the evaluation of each batch of individuals. Finally, a real-world engineering application that consists on optimizing the refrigeration system of a power electronic device has been presented as an illustrative example in which the use of the proposed load balancing algorithms is able to reduce the simulation time required by the optimization tool.El aumento de las aplicaciones que requieren de una aproximación multidisciplinar para poder avanzar se constata en todos los campos de la ingeniería, lo cual conlleva la necesidad de resolver problemas de optimización complejos que exceden la capacidad del cerebro humano o de la intuición. En estos casos es habitual el uso de algoritmos evolutivos, principalmente de los algoritmos genéticos, caracterizados por su robustez y versatilidad, así como por su gran coste computacional y baja velocidad de convergencia. La multitud de paquetes de optimización disponibles con licencias de software libre representan el estado del arte actual en tecnología de optimización. Sin embargo, la capacidad de adaptación de los algoritmos de optimización a ordenadores masivamente paralelos alcanzando niveles de eficiencia satisfactorios es todavía una tarea pendiente. Incluso los paquetes adaptados al paralelismo multinivel tienen dificultades para gestionar funciones objetivo que requieren de tiempos de simulación largos y variables. Esta variabilidad es común en la Dinámica de Fluidos Computacional y la Transferencia de Calor (CFD & HT), mecánica no lineal, etc. y es una de las principales preocupaciones en aplicaciones a gran escala a día de hoy. La investigación actual que tiene por objetivo la mejora del rendimiento de los algoritmos evolutivos está enfocada principalmente al desarrollo de nuevos algoritmos de búsqueda. Sin embargo, ya se conoce una gran variedad de algoritmos secuenciales apropiados para su implementación en ordenadores paralelos. La tarea pendiente es conseguir una paralelización eficiente. Además, los avances en la investigación de nuevos algoritmos de búsqueda y la paralelización son aditivos, por lo que el proceso de mejora del software de optimización actual se verá incrementada si se atacan ambos frentes simultáneamente. La motivación de esta Tesis Doctoral es avanzar hacia una integración completa de las capacidades de Optimización y Computación de Alto Rendimiento para así impulsar el desarrollo tecnológico proporcionando mejores diseños, acortando los tiempos de desarrollo del producto y minimizando los recursos necesarios. Tras un exhaustivo estudio del estado del arte de las técnicas de optimización matemática disponibles a día de hoy, se ha diseñado una librería de optimización orientada al campo de la Dinámica de Fluidos Computacional y la Transferencia de Calor (CFD & HT). A continuación se han analizado las principales limitaciones de las estrategias de paralelización disponibles para algoritmos genéticos y otros métodos de optimización basados en poblaciones. En el caso en que el tiempo de evaluación medio de la tanda de individuos sea mayor que el tiempo medio que necesita el optimizador para llevar a cabo comunicaciones entre procesadores, se ha detectado que la causa principal de la degradación de la escalabilidad o eficiencia paralela del algoritmo de optimización es el desequilibrio de la carga computacional. El motivo es que a menudo los procesadores no terminan de evaluar su cola de individuos simultáneamente y deben sincronizarse antes de que se cree la siguiente tanda de individuos. Por consiguiente, el desequilibrio de la carga computacional se convierte en tiempo de inactividad en algunos procesadores. Se han propuesto y testado exhaustivamente varios algoritmos de equilibrado de carga aplicables a cualquier método de optimización basado en una población que necesite sincronizar los procesadores tras cada tanda de evaluaciones. Finalmente, se ha presentado como ejemplo ilustrativo un caso real de ingeniería que consiste en optimizar el sistema de refrigeración de un dispositivo de electrónica de potencia. En él queda demostrado que el uso de los algoritmos de equilibrado de carga computacional propuestos es capaz de reducir el tiempo de simulación que necesita la herramienta de optimización
    • …
    corecore