6,569 research outputs found
State-of-the-art in aerodynamic shape optimisation methods
Aerodynamic optimisation has become an indispensable component for any aerodynamic design over the past 60 years, with applications to aircraft, cars, trains, bridges, wind turbines, internal pipe flows, and cavities, among others, and is thus relevant in many facets of technology. With advancements in computational power, automated design optimisation procedures have become more competent, however, there is an ambiguity and bias throughout the literature with regards to relative performance of optimisation architectures and employed algorithms. This paper provides a well-balanced critical review of the dominant optimisation approaches that have been integrated with aerodynamic theory for the purpose of shape optimisation. A total of 229 papers, published in more than 120 journals and conference proceedings, have been classified into 6 different optimisation algorithm approaches. The material cited includes some of the most well-established authors and publications in the field of aerodynamic optimisation. This paper aims to eliminate bias toward certain algorithms by analysing the limitations, drawbacks, and the benefits of the most utilised optimisation approaches. This review provides comprehensive but straightforward insight for non-specialists and reference detailing the current state for specialist practitioners
Enhanced genetic algorithm-based fuzzy multiobjective strategy to multiproduct batch plant design
This paper addresses the problem of the optimal design of batch plants with imprecise demands in product amounts. The design of such plants necessary involves how equipment may be utilized, which means that plant scheduling and production must constitute a basic part of the design problem. Rather than resorting to a traditional probabilistic approach for modeling the imprecision on product demands, this work proposes an alternative treatment by using fuzzy concepts. The design problem is tackled by introducing a new approach based on a multiobjective genetic algorithm, combined wit the fuzzy set theory for computing the objectives as fuzzy quantities. The problem takes into account simultaneous maximization of the fuzzy net present value and of two other performance criteria, i.e. the production delay/advance and a flexibility index. The delay/advance objective is computed by comparing the fuzzy production time for the products to a given fuzzy time horizon, and the flexibility index represents the additional fuzzy production that the plant would be able to produce. The multiobjective optimization provides the Pareto's front which is a set of scenarios that are helpful for guiding the decision's maker in its final choices. About the solution procedure, a genetic algorithm was implemented since it is particularly well-suited to take into account the arithmetic of fuzzy numbers. Furthermore because a genetic algorithm is working on populations of potential solutions, this type of procedure is well adapted for multiobjective optimization
Stochastic make-to-stock inventory deployment problem: an endosymbiotic psychoclonal algorithm based approach
Integrated steel manufacturers (ISMs) have no specific product, they just produce finished product from the ore. This enhances the uncertainty prevailing in the ISM regarding the nature of the finished product and significant demand by customers. At present low cost mini-mills are giving firm competition to ISMs in terms of cost, and this has compelled the ISM industry to target customers who want exotic products and faster reliable deliveries. To meet this objective, ISMs are exploring the option of satisfying part of their demand by converting strategically placed products, this helps in increasing the variability of product produced by the ISM in a short lead time. In this paper the authors have proposed a new hybrid evolutionary algorithm named endosymbiotic-psychoclonal (ESPC) to decide what and how much to stock as a semi-product in inventory. In the proposed theory, the ability of previously proposed psychoclonal algorithms to exploit the search space has been increased by making antibodies and antigen more co-operative interacting species. The efficacy of the proposed algorithm has been tested on randomly generated datasets and the results compared with other evolutionary algorithms such as genetic algorithms (GA) and simulated annealing (SA). The comparison of ESPC with GA and SA proves the superiority of the proposed algorithm both in terms of quality of the solution obtained and convergence time required to reach the optimal/near optimal value of the solution
An ESPC algorithm based approach to solve inventory deployment problem
Global competitiveness has enforced the hefty industries to become more customized. To compete in the market they are targeting the customers who want exotic products, and faster and reliable deliveries. Industries are exploring the option of satisfying a portion of their demand by converting strategically placed products, this helps in increasing the variability of product produced by them in short lead time. In this paper, authors have proposed a new hybrid evolutionary algorithm named Endosymbiotic-Psychoclonal (ESPC) algorithm to determine the amount and type of product to stock as a semi product in inventory. In the proposed work the ability of previously proposed Psychoclonal algorithm to exploit the search space has been increased by making antibodies and antigen more cooperative interacting species. The efficacy of the proposed algorithm has been tested on randomly generated datasets and the results obtained, are compared with other evolutionary algorithms such as Genetic Algorithm (GA) and Simulated Annealing (SA). The comparison of ESPC with GA and SA proves the superiority of the proposed algorithm both in terms of quality of the solution obtained, and convergence time required to reach the optimal /near optimal value of the solution
Recommended from our members
Local search: A guide for the information retrieval practitioner
There are a number of combinatorial optimisation problems in information retrieval in which the use of local search methods are worthwhile. The purpose of this paper is to show how local search can be used to solve some well known tasks in information retrieval (IR), how previous research in the field is piecemeal, bereft of a structure and methodologically flawed, and to suggest more rigorous ways of applying local search methods to solve IR problems. We provide a query based taxonomy for analysing the use of local search in IR tasks and an overview of issues such as fitness functions, statistical significance and test collections when conducting experiments on combinatorial optimisation problems. The paper gives a guide on the pitfalls and problems for IR practitioners who wish to use local search to solve their research issues, and gives practical advice on the use of such methods. The query based taxonomy is a novel structure which can be used by the IR practitioner in order to examine the use of local search in IR
Review of Metaheuristics and Generalized Evolutionary Walk Algorithm
Metaheuristic algorithms are often nature-inspired, and they are becoming
very powerful in solving global optimization problems. More than a dozen of
major metaheuristic algorithms have been developed over the last three decades,
and there exist even more variants and hybrid of metaheuristics. This paper
intends to provide an overview of nature-inspired metaheuristic algorithms,
from a brief history to their applications. We try to analyze the main
components of these algorithms and how and why they works. Then, we intend to
provide a unified view of metaheuristics by proposing a generalized
evolutionary walk algorithm (GEWA). Finally, we discuss some of the important
open questions.Comment: 14 page
Recommended from our members
Combinatorial optimization and metaheuristics
Today, combinatorial optimization is one of the youngest and most active areas of discrete mathematics. It is a branch of optimization in applied mathematics and computer science, related to operational research, algorithm theory and computational complexity theory. It sits at the intersection of several fields, including artificial intelligence, mathematics and software engineering. Its increasing interest arises for the fact that a large number of scientific and industrial problems can be formulated as abstract combinatorial optimization problems, through graphs and/or (integer) linear programs. Some of these problems have polynomial-time (“efficient”) algorithms, while most of them are NP-hard, i.e. it is not proved that they can be solved in polynomial-time. Mainly, it means that it is not possible to guarantee that an exact solution to the problem can be found and one has to settle for an approximate solution with known performance guarantees. Indeed, the goal of approximate methods is to find “quickly” (reasonable run-times), with “high” probability, provable “good” solutions (low error from the real optimal solution). In the last 20 years, a new kind of algorithm commonly called metaheuristics have emerged in this class, which basically try to combine heuristics in high level frameworks aimed at efficiently and effectively exploring the search space. This report briefly outlines the components, concepts, advantages and disadvantages of different metaheuristic approaches from a conceptual point of view, in order to analyze their similarities and differences. The two very significant forces of intensification and diversification, that mainly determine the behavior of a metaheuristic, will be pointed out. The report concludes by exploring the importance of hybridization and integration methods
Using Artificial Intelligence for Model Selection
We apply the optimization algorithm Adaptive Simulated Annealing (ASA) to the
problem of analyzing data on a large population and selecting the best model to
predict that an individual with various traits will have a particular disease.
We compare ASA with traditional forward and backward regression on computer
simulated data. We find that the traditional methods of modeling are better for
smaller data sets whereas a numerically stable ASA seems to perform better on
larger and more complicated data sets.Comment: 10 pages, no figures, in Proceedings, Hawaii International Conference
on Statistics and Related Fields, June 5-8, 200
- …