25 research outputs found
Multi-criteria Optimization of Workflow-Based Assembly Tasks in Manufacturing
Industrial manufacturing is currently amidst it's fourth great revolution,
pushing towards the digital transformation of production processes. One key
element of this transformation is the formalization and digitization of
processes, creating an increased potential to monitor, understand and optimize
existing processes. However, one major obstacle in this process is the
increased diversification and specialisation, resulting in the dependency on
multiple experts, which are rarely amalgamated in small to medium sized
companies. To mitigate this issue, this paper presents a novel approach for
multi-criteria optimization of workflow-based assembly tasks in manufacturing
by combining a workflow modeling framework and the HeuristicLab optimization
framework. For this endeavour, a new generic problem definition is implemented
in HeuristicLab, enabling the optimization of arbitrary workflows represented
with the modeling framework. The resulting Pareto front of the multi-criteria
optimization provides the decision makers a set of optimal workflows from which
they can choose to optimally fit the current demands. The advantages of the
herein presented approach are highlighted with a real world use case from an
ongoing research project.Comment: This preprint has not undergone peer review or any post-submission
improvements or corrections. The Version of Record of this contribution is
published in Computer Aided Systems Theory - EUROCAST 2022 and is available
online at this URL https://doi.org/10.1007/978-3-031-25312-6_
Optimizing the performance of optimization in the cloud environment–An intelligent auto-scaling approach
The cloud computing paradigm has gained wide acceptance in the scientific community, taking a significant share from fields previously reserved exclusively for High Performance Computing (HPC). On-demand access to a large amount of computing resources provided by Cloud makes it ideal for executing large-scale optimizations using evolutionary algorithms without the need for owning any computing infrastructure. In this regard, we extended WoBinGO, an existing parallel software framework for genetic algorithm based optimization, to be used in Cloud. With these extensions, the framework is capable of elastically and frugally utilizing the underlying cloud computing infrastructure for performing computationally expensive fitness evaluations.
We studied two issues that are pertinent when dealing with large-scale optimization in the elastic cloud environment: the computing instance launching overhead and the price of engaging Cloud for solving optimization problems, in terms of the instances’ cumulative uptime. To explain the usability limits of WoBinGO framework running in the IaaS environment, a comprehensive analysis of the framework’s performance was given.
Optimization of both total optimization time and total cumulative uptime, leads to minimizing the cost of cloud resources utilization. In this way, we are proposing an intelligent decision support engine based on artificial neural networks and metaheuristics to provide the user with an assessment of the framework’s behavior on the underlying infrastructure in terms of optimization duration and the cost of resource consumption. According to a given assessment, the user can decide upon faster delivery of results or lower infrastructure costs.
The proposed software framework has been used to solve a complex real-world optimization problem of a subsurface rock mass model calibration. The results obtained from the private OpenStack deployment show that by using the proposed decision support engine, significant savings can be achieved in both optimization time and optimization cost
Optimizing physical protection system using domain experienced exploration method
Assessing physical protection system efficiency is mostly done manually by security experts due to the complexity of the assessment process and lack of tools. Computer aided numerical vulnerability analysis has been developed to quantitatively assess physical protection systems. A variety of methods have been proposed to optimize physical protection systems, where one of the most advanced approaches entails precisely defining which security components should be selected and where they should be placed at protected facilities, taking into consideration adversary type, to maximize the probability that an adversary will be stopped at minimum system cost. The most computationally intensive part of the optimization process is the evaluation. The evaluation involves recreating search space and finding optimal adversary’s attack paths from each entry point. We present the domain experienced exploration method that optimizes evaluation process during the search for optimum solutions, considering results from previous evaluations. Performed experiments show that using the presented method, in real-world domains, results in a reduction of evaluation iterations
Recommended from our members
OptPlatform: metaheuristic optimisation framework for solving complex real-world problems
This thesis was submitted for the award of Doctor of Philosophy and was awarded by Brunel University LondonWe optimise daily, whether that is planning a round trip that visits the most attractions within a given holiday budget or just taking a train instead of driving a car in a rush hour. Many problems, just like these, are solved by individuals as part of our daily schedule, and they are effortless and straightforward. If we now scale that to many individuals with many different schedules, like a school timetable, we get to a point where it is just not feasible or practical to solve by hand. In such instances, optimisation methods are used to obtain an optimal solution. In this thesis, a practical approach to optimisation has been taken by developing an optimisation platform with all the necessary tools to be used by practitioners who are not necessarily familiar with the subject of optimisation. First, a high-performance metaheuristic optimisation framework (MOF) called OptPlatform is implemented, and the versatility and performance are evaluated across multiple benchmarks and real-world optimisation problems. Results show that, compared to competing MOFs, the OptPlatform outperforms in both the solution quality and computation time. Second, the most suitable hardware platform for OptPlatform is determined by an in-depth analysis of Ant Colony Optimisation scaling across CPU, GPU and enterprise Xeon Phi. Contrary to the common benchmark problems used in the literature, the supply chain problem solved could not scale on GPUs. Third, a variety of metaheuristics are implemented into OptPlatform. Including, a new metaheuristic based on Imperialist Competitive Algorithm (ICA), called ICA with Independence and Constrained Assimilation (ICAwICA) is proposed. The ICAwICA was compared against two different types of benchmark problems, and results show the versatile application of the algorithm, matching and in some cases outperforming the custom-tuned approaches. Finally, essential MOF features like automatic algorithm selection and tuning, lacking on existing frameworks, are implemented in OptPlatform. Two novel approaches are proposed and compared to existing methods. Results indicate the superiority of the implemented tuning algorithms within constrained tuning budget environment
Local Search is Underused in Genetic Programming
Trujillo, L., Z-Flores, E., Juárez-Smith, P. S., Legrand, P., Silva, S., Castelli, M., ... Muñoz, L. (2018). Local Search is Underused in Genetic Programming. In R. Riolo, B. Worzel, B. Goldman, & B. Tozier (Eds.), Genetic Programming Theory and Practice XIV (pp. 119-137). [8] (Genetic and Evolutionary Computation). Springer. https://doi.org/10.1007/978-3-319-97088-2_8There are two important limitations of standard tree-based genetic programming (GP). First, GP tends to evolve unnecessarily large programs, what is referred to as bloat. Second, GP uses inefficient search operators that focus on modifying program syntax. The first problem has been studied extensively, with many works proposing bloat control methods. Regarding the second problem, one approach is to use alternative search operators, for instance geometric semantic operators, to improve convergence. In this work, our goal is to experimentally show that both problems can be effectively addressed by incorporating a local search optimizer as an additional search operator. Using real-world problems, we show that this rather simple strategy can improve the convergence and performance of tree-based GP, while also reducing program size. Given these results, a question arises: Why are local search strategies so uncommon in GP? A small survey of popular GP libraries suggests to us that local search is underused in GP systems. We conclude by outlining plausible answers for this question and highlighting future work.authorsversionpublishe
Optimization and Prediction Techniques for Self-Healing and Self-Learning Applications in a Trustworthy Cloud Continuum
The current IT market is more and more dominated by the “cloud continuum”. In the “traditional” cloud, computing resources are typically homogeneous in order to facilitate economies of scale. In contrast, in edge computing, computational resources are widely diverse, commonly with scarce capacities and must be managed very efficiently due to battery constraints or other limitations. A combination of resources and services at the edge (edge computing), in the core (cloud computing), and along the data path (fog computing) is needed through a trusted cloud continuum. This requires novel solutions for the creation, optimization, management, and automatic operation of such infrastructure through new approaches such as infrastructure as code (IaC). In this paper, we analyze how artificial intelligence (AI)-based techniques and tools can enhance the operation of complex applications to support the broad and multi-stage heterogeneity of the infrastructural layer in the “computing continuum” through the enhancement of IaC optimization, IaC self-learning, and IaC self-healing. To this extent, the presented work proposes a set of tools, methods, and techniques for applications’ operators to seamlessly select, combine, configure, and adapt computation resources all along the data path and support the complete service lifecycle covering: (1) optimized distributed application deployment over heterogeneous computing resources; (2) monitoring of execution platforms in real time including continuous control and trust of the infrastructural services; (3) application deployment and adaptation while optimizing the execution; and (4) application self-recovery to avoid compromising situations that may lead to an unexpected failure.This research was funded by the European project PIACERE (Horizon 2020 research and innovation Program, under grant agreement no 101000162)
Multi-objective tools for the vehicle routing problem with time windows
Most real-life problems involve the simultaneous optimisation of two or more, usually conflicting, objectives. Researchers have put a continuous effort into solving these problems in many different areas, such as engineering, finance and computer science. Over time, thanks to the increase in processing power, researchers have created methods which have become increasingly sophisticated. Most of these methods have been based on the notion of Pareto dominance, which assumes, sometimes erroneously, that the objectives have no known ranking of importance.
The Vehicle Routing Problem with Time Windows (VRPTW) is a logistics problem which in real-life applications appears to be multi-objective. This problem consists of designing the optimal set of routes to serve a number of customers within certain time slots. Despite this problem’s high applicability to real-life domains (e.g. waste collection, fast-food delivery), most research in this area has been conducted with hand-made datasets. These datasets sometimes have a number of unrealistic features (e.g. the assumption that one unit of travel time corresponds to one unit of travel distance) and are therefore not adequate for the assessment of optimisers. Furthermore, very few studies have focused on the multi-objective nature of the VRPTW. That is, very few have studied how the optimisation of one objective affects the others.
This thesis proposes a number of novel tools (methods + dataset) to address the above- mentioned challenges: 1) an agent-based framework for cooperative search, 2) a novel multi-objective ranking approach, 3) a new dataset for the VRPTW, 4) a study of the pair-wise relationships between five common objectives in VRPTW, and 5) a simplified Multi-objective Discrete Particle Swarm Optimisation for the VRPTW
Multi-objective tools for the vehicle routing problem with time windows
Most real-life problems involve the simultaneous optimisation of two or more, usually conflicting, objectives. Researchers have put a continuous effort into solving these problems in many different areas, such as engineering, finance and computer science. Over time, thanks to the increase in processing power, researchers have created methods which have become increasingly sophisticated. Most of these methods have been based on the notion of Pareto dominance, which assumes, sometimes erroneously, that the objectives have no known ranking of importance.
The Vehicle Routing Problem with Time Windows (VRPTW) is a logistics problem which in real-life applications appears to be multi-objective. This problem consists of designing the optimal set of routes to serve a number of customers within certain time slots. Despite this problem’s high applicability to real-life domains (e.g. waste collection, fast-food delivery), most research in this area has been conducted with hand-made datasets. These datasets sometimes have a number of unrealistic features (e.g. the assumption that one unit of travel time corresponds to one unit of travel distance) and are therefore not adequate for the assessment of optimisers. Furthermore, very few studies have focused on the multi-objective nature of the VRPTW. That is, very few have studied how the optimisation of one objective affects the others.
This thesis proposes a number of novel tools (methods + dataset) to address the above- mentioned challenges: 1) an agent-based framework for cooperative search, 2) a novel multi-objective ranking approach, 3) a new dataset for the VRPTW, 4) a study of the pair-wise relationships between five common objectives in VRPTW, and 5) a simplified Multi-objective Discrete Particle Swarm Optimisation for the VRPTW
Population-based simulation optimization for urban mass rapid transit networks
In this paper, we present a simulation-based headway optimization for urban mass rapid transit networks. The underlying discrete event simulation model contains several stochastic elements, including time-dependent demand and turning maneuver times as well as direction-dependent vehicle travel and passenger transfer times. Passenger creation is a Poisson process that uses hourly origin–destination-matrices based on anonymous mobile phone and infrared count data. The numbers of passengers on platforms and within vehicles are subject to capacity restrictions. As a microscopic element, passenger distribution along platforms and within vehicles is considered. The bi-objective problem, involving cost reduction and service level improvement, is transformed into a single-objective optimization problem by normalization and scalarization. Population-based evolutionary algorithms and different solution encoding variants are applied. Computational experience is gained from test instances based on real-world data (i.e., the Viennese subway network). A covariance matrix adaptation evolution strategy performs best in most cases, and a newly developed encoding helps accelerate the optimization process by producing better short-term results.
Document type: Articl