25 research outputs found

    Multi-criteria Optimization of Workflow-Based Assembly Tasks in Manufacturing

    Full text link
    Industrial manufacturing is currently amidst it's fourth great revolution, pushing towards the digital transformation of production processes. One key element of this transformation is the formalization and digitization of processes, creating an increased potential to monitor, understand and optimize existing processes. However, one major obstacle in this process is the increased diversification and specialisation, resulting in the dependency on multiple experts, which are rarely amalgamated in small to medium sized companies. To mitigate this issue, this paper presents a novel approach for multi-criteria optimization of workflow-based assembly tasks in manufacturing by combining a workflow modeling framework and the HeuristicLab optimization framework. For this endeavour, a new generic problem definition is implemented in HeuristicLab, enabling the optimization of arbitrary workflows represented with the modeling framework. The resulting Pareto front of the multi-criteria optimization provides the decision makers a set of optimal workflows from which they can choose to optimally fit the current demands. The advantages of the herein presented approach are highlighted with a real world use case from an ongoing research project.Comment: This preprint has not undergone peer review or any post-submission improvements or corrections. The Version of Record of this contribution is published in Computer Aided Systems Theory - EUROCAST 2022 and is available online at this URL https://doi.org/10.1007/978-3-031-25312-6_

    Optimizing the performance of optimization in the cloud environment–An intelligent auto-scaling approach

    Get PDF
    The cloud computing paradigm has gained wide acceptance in the scientific community, taking a significant share from fields previously reserved exclusively for High Performance Computing (HPC). On-demand access to a large amount of computing resources provided by Cloud makes it ideal for executing large-scale optimizations using evolutionary algorithms without the need for owning any computing infrastructure. In this regard, we extended WoBinGO, an existing parallel software framework for genetic algorithm based optimization, to be used in Cloud. With these extensions, the framework is capable of elastically and frugally utilizing the underlying cloud computing infrastructure for performing computationally expensive fitness evaluations. We studied two issues that are pertinent when dealing with large-scale optimization in the elastic cloud environment: the computing instance launching overhead and the price of engaging Cloud for solving optimization problems, in terms of the instances’ cumulative uptime. To explain the usability limits of WoBinGO framework running in the IaaS environment, a comprehensive analysis of the framework’s performance was given. Optimization of both total optimization time and total cumulative uptime, leads to minimizing the cost of cloud resources utilization. In this way, we are proposing an intelligent decision support engine based on artificial neural networks and metaheuristics to provide the user with an assessment of the framework’s behavior on the underlying infrastructure in terms of optimization duration and the cost of resource consumption. According to a given assessment, the user can decide upon faster delivery of results or lower infrastructure costs. The proposed software framework has been used to solve a complex real-world optimization problem of a subsurface rock mass model calibration. The results obtained from the private OpenStack deployment show that by using the proposed decision support engine, significant savings can be achieved in both optimization time and optimization cost

    Optimizing physical protection system using domain experienced exploration method

    Get PDF
    Assessing physical protection system efficiency is mostly done manually by security experts due to the complexity of the assessment process and lack of tools. Computer aided numerical vulnerability analysis has been developed to quantitatively assess physical protection systems. A variety of methods have been proposed to optimize physical protection systems, where one of the most advanced approaches entails precisely defining which security components should be selected and where they should be placed at protected facilities, taking into consideration adversary type, to maximize the probability that an adversary will be stopped at minimum system cost. The most computationally intensive part of the optimization process is the evaluation. The evaluation involves recreating search space and finding optimal adversary’s attack paths from each entry point. We present the domain experienced exploration method that optimizes evaluation process during the search for optimum solutions, considering results from previous evaluations. Performed experiments show that using the presented method, in real-world domains, results in a reduction of evaluation iterations

    Local Search is Underused in Genetic Programming

    Get PDF
    Trujillo, L., Z-Flores, E., Juárez-Smith, P. S., Legrand, P., Silva, S., Castelli, M., ... Muñoz, L. (2018). Local Search is Underused in Genetic Programming. In R. Riolo, B. Worzel, B. Goldman, & B. Tozier (Eds.), Genetic Programming Theory and Practice XIV (pp. 119-137). [8] (Genetic and Evolutionary Computation). Springer. https://doi.org/10.1007/978-3-319-97088-2_8There are two important limitations of standard tree-based genetic programming (GP). First, GP tends to evolve unnecessarily large programs, what is referred to as bloat. Second, GP uses inefficient search operators that focus on modifying program syntax. The first problem has been studied extensively, with many works proposing bloat control methods. Regarding the second problem, one approach is to use alternative search operators, for instance geometric semantic operators, to improve convergence. In this work, our goal is to experimentally show that both problems can be effectively addressed by incorporating a local search optimizer as an additional search operator. Using real-world problems, we show that this rather simple strategy can improve the convergence and performance of tree-based GP, while also reducing program size. Given these results, a question arises: Why are local search strategies so uncommon in GP? A small survey of popular GP libraries suggests to us that local search is underused in GP systems. We conclude by outlining plausible answers for this question and highlighting future work.authorsversionpublishe

    Optimization and Prediction Techniques for Self-Healing and Self-Learning Applications in a Trustworthy Cloud Continuum

    Get PDF
    The current IT market is more and more dominated by the “cloud continuum”. In the “traditional” cloud, computing resources are typically homogeneous in order to facilitate economies of scale. In contrast, in edge computing, computational resources are widely diverse, commonly with scarce capacities and must be managed very efficiently due to battery constraints or other limitations. A combination of resources and services at the edge (edge computing), in the core (cloud computing), and along the data path (fog computing) is needed through a trusted cloud continuum. This requires novel solutions for the creation, optimization, management, and automatic operation of such infrastructure through new approaches such as infrastructure as code (IaC). In this paper, we analyze how artificial intelligence (AI)-based techniques and tools can enhance the operation of complex applications to support the broad and multi-stage heterogeneity of the infrastructural layer in the “computing continuum” through the enhancement of IaC optimization, IaC self-learning, and IaC self-healing. To this extent, the presented work proposes a set of tools, methods, and techniques for applications’ operators to seamlessly select, combine, configure, and adapt computation resources all along the data path and support the complete service lifecycle covering: (1) optimized distributed application deployment over heterogeneous computing resources; (2) monitoring of execution platforms in real time including continuous control and trust of the infrastructural services; (3) application deployment and adaptation while optimizing the execution; and (4) application self-recovery to avoid compromising situations that may lead to an unexpected failure.This research was funded by the European project PIACERE (Horizon 2020 research and innovation Program, under grant agreement no 101000162)

    Multi-objective tools for the vehicle routing problem with time windows

    Get PDF
    Most real-life problems involve the simultaneous optimisation of two or more, usually conflicting, objectives. Researchers have put a continuous effort into solving these problems in many different areas, such as engineering, finance and computer science. Over time, thanks to the increase in processing power, researchers have created methods which have become increasingly sophisticated. Most of these methods have been based on the notion of Pareto dominance, which assumes, sometimes erroneously, that the objectives have no known ranking of importance. The Vehicle Routing Problem with Time Windows (VRPTW) is a logistics problem which in real-life applications appears to be multi-objective. This problem consists of designing the optimal set of routes to serve a number of customers within certain time slots. Despite this problem’s high applicability to real-life domains (e.g. waste collection, fast-food delivery), most research in this area has been conducted with hand-made datasets. These datasets sometimes have a number of unrealistic features (e.g. the assumption that one unit of travel time corresponds to one unit of travel distance) and are therefore not adequate for the assessment of optimisers. Furthermore, very few studies have focused on the multi-objective nature of the VRPTW. That is, very few have studied how the optimisation of one objective affects the others. This thesis proposes a number of novel tools (methods + dataset) to address the above- mentioned challenges: 1) an agent-based framework for cooperative search, 2) a novel multi-objective ranking approach, 3) a new dataset for the VRPTW, 4) a study of the pair-wise relationships between five common objectives in VRPTW, and 5) a simplified Multi-objective Discrete Particle Swarm Optimisation for the VRPTW

    Multi-objective tools for the vehicle routing problem with time windows

    Get PDF
    Most real-life problems involve the simultaneous optimisation of two or more, usually conflicting, objectives. Researchers have put a continuous effort into solving these problems in many different areas, such as engineering, finance and computer science. Over time, thanks to the increase in processing power, researchers have created methods which have become increasingly sophisticated. Most of these methods have been based on the notion of Pareto dominance, which assumes, sometimes erroneously, that the objectives have no known ranking of importance. The Vehicle Routing Problem with Time Windows (VRPTW) is a logistics problem which in real-life applications appears to be multi-objective. This problem consists of designing the optimal set of routes to serve a number of customers within certain time slots. Despite this problem’s high applicability to real-life domains (e.g. waste collection, fast-food delivery), most research in this area has been conducted with hand-made datasets. These datasets sometimes have a number of unrealistic features (e.g. the assumption that one unit of travel time corresponds to one unit of travel distance) and are therefore not adequate for the assessment of optimisers. Furthermore, very few studies have focused on the multi-objective nature of the VRPTW. That is, very few have studied how the optimisation of one objective affects the others. This thesis proposes a number of novel tools (methods + dataset) to address the above- mentioned challenges: 1) an agent-based framework for cooperative search, 2) a novel multi-objective ranking approach, 3) a new dataset for the VRPTW, 4) a study of the pair-wise relationships between five common objectives in VRPTW, and 5) a simplified Multi-objective Discrete Particle Swarm Optimisation for the VRPTW

    Population-based simulation optimization for urban mass rapid transit networks

    Get PDF
    In this paper, we present a simulation-based headway optimization for urban mass rapid transit networks. The underlying discrete event simulation model contains several stochastic elements, including time-dependent demand and turning maneuver times as well as direction-dependent vehicle travel and passenger transfer times. Passenger creation is a Poisson process that uses hourly origin–destination-matrices based on anonymous mobile phone and infrared count data. The numbers of passengers on platforms and within vehicles are subject to capacity restrictions. As a microscopic element, passenger distribution along platforms and within vehicles is considered. The bi-objective problem, involving cost reduction and service level improvement, is transformed into a single-objective optimization problem by normalization and scalarization. Population-based evolutionary algorithms and different solution encoding variants are applied. Computational experience is gained from test instances based on real-world data (i.e., the Viennese subway network). A covariance matrix adaptation evolution strategy performs best in most cases, and a newly developed encoding helps accelerate the optimization process by producing better short-term results. Document type: Articl
    corecore