2,693 research outputs found
State-of-the-art in aerodynamic shape optimisation methods
Aerodynamic optimisation has become an indispensable component for any aerodynamic design over the past 60 years, with applications to aircraft, cars, trains, bridges, wind turbines, internal pipe flows, and cavities, among others, and is thus relevant in many facets of technology. With advancements in computational power, automated design optimisation procedures have become more competent, however, there is an ambiguity and bias throughout the literature with regards to relative performance of optimisation architectures and employed algorithms. This paper provides a well-balanced critical review of the dominant optimisation approaches that have been integrated with aerodynamic theory for the purpose of shape optimisation. A total of 229 papers, published in more than 120 journals and conference proceedings, have been classified into 6 different optimisation algorithm approaches. The material cited includes some of the most well-established authors and publications in the field of aerodynamic optimisation. This paper aims to eliminate bias toward certain algorithms by analysing the limitations, drawbacks, and the benefits of the most utilised optimisation approaches. This review provides comprehensive but straightforward insight for non-specialists and reference detailing the current state for specialist practitioners
A hybrid global optimization method: The multi-dimensional case
AbstractWe extend the hybrid global optimization method proposed by Xu (J. Comput. Appl. Math. 147 (2002) 301–314) for the one-dimensional case to the multi-dimensional case. The method consists of two basic components: local optimizers and feasible point finders. Local optimizers guarantee efficiency and speed of producing a local optimal solution in the neighbourhood of a feasible point. Feasible point finders provide the theoretical guarantee for the new method to always produce the global optimal solution(s) correctly. If a nonlinear nonconvex inverse problem has multiple global optimal solutions, our algorithm is capable of finding all of them correctly. Three synthetic examples, which have failed simulated annealing and genetic algorithms, are used to demonstrate the proposed method
The design and applications of the african buffalo algorithm for general optimization problems
Optimization, basically, is the economics of science. It is concerned with the need to maximize profit and minimize cost in terms of time and resources needed to execute a given project in any field of human endeavor. There have been several scientific investigations in the past several decades on discovering effective and efficient algorithms to providing solutions to the optimization needs of mankind leading to the development
of deterministic algorithms that provide exact solutions to optimization problems. In the past five decades, however, the attention of scientists has shifted from the deterministic algorithms to the stochastic ones since the latter have proven to be more robust and efficient, even though they do not guarantee exact solutions. Some of the successfully designed stochastic algorithms include Simulated Annealing, Genetic Algorithm, Ant Colony Optimization, Particle Swarm Optimization, Bee Colony Optimization, Artificial Bee Colony Optimization, Firefly Optimization etc. A critical look at these ‘efficient’
stochastic algorithms reveals the need for improvements in the areas of effectiveness, the number of several parameters used, premature convergence, ability to search diverse landscapes and complex implementation strategies. The African Buffalo Optimization (ABO), which is inspired by the herd management, communication and successful
grazing cultures of the African buffalos, is designed to attempt solutions to the observed shortcomings of the existing stochastic optimization algorithms. Through several experimental procedures, the ABO was used to successfully solve benchmark optimization problems in mono-modal and multimodal, constrained and unconstrained,
separable and non-separable search landscapes with competitive outcomes. Moreover, the ABO algorithm was applied to solve over 100 out of the 118 benchmark symmetric and all the asymmetric travelling salesman’s problems available in TSPLIB95. Based on the
successful experimentation with the novel algorithm, it is safe to conclude that the ABO is a worthy contribution to the scientific literature
Adaptive Penalty and Barrier function based on Fuzzy Logic
Optimization methods have been used in many areas of knowledge, such as Engineering, Statistics, Chemistry, among others, to solve optimization problems. In many cases it is not possible to use derivative methods, due to the characteristics of the problem to be solved and/or its constraints, for example if the involved functions are non-smooth and/or their derivatives are not know. To solve this type of problems a Java based API has been implemented, which includes only derivative-free optimization methods, and that can be used to solve both constrained and unconstrained problems. For solving constrained problems, the classic Penalty and Barrier functions were included in the API. In this paper a new approach to Penalty and Barrier functions, based on Fuzzy Logic, is proposed. Two penalty functions, that impose a progressive penalization to solutions that violate the constraints, are discussed. The implemented functions impose a low penalization when the violation of the constraints is low and a heavy penalty when the violation is high. Numerical results, obtained using twenty-eight test problems, comparing the proposed Fuzzy Logic based functions to six of the classic Penalty and Barrier functions are presented. Considering the achieved results, it can be concluded that the proposed penalty functions besides being very robust also have a very good performance
Dynamic Optimization on Quantum Hardware: Feasibility for a Process Industry Use Case
The quest for real-time dynamic optimization solutions in the process
industry represents a formidable computational challenge, particularly within
the realm of applications like model predictive control where rapid and
reliable computations are critical. Conventional methods can struggle to
surmount the complexities of such tasks. Quantum computing and quantum
annealing emerge as avant-garde contenders to transcend conventional
computational constraints. We convert a dynamic optimization problem,
characterized by a system of differential equations, into a Quadratic
Unconstrained Binary Optimization problem, enabling quantum computational
approaches. The empirical findings synthesized from classical methods,
simulated annealing, quantum annealing via D-Wave's quantum annealer, and
hybrid solver methodologies, illuminate the intricate landscape of
computational prowess essential for tackling complex and high-dimensional
dynamic optimization problems. Our findings suggest that while quantum
annealing is a maturing technology that currently does not outperform
state-of-the-art classical solvers, continuous improvements could eventually
aid in increasing efficiency within the chemical process industry.Comment: 17 pages, 5 figure
A Family of Hybrid Stochastic Conjugate Gradient Algorithms for Local and Global Minimization Problems
This paper contains two main parts, Part I and Part II, which discuss the local and global minimization problems, respectively. In Part I, a fresh conjugate gradient (CG) technique is suggested and then combined with a line-search technique to obtain a globally convergent algorithm. The finite difference approximations approach is used to compute the approximate values of the first derivative of the function f. The convergence analysis of the suggested method is established. The comparisons between the performance of the new CG method and the performance of four other CG methods demonstrate that the proposed CG method is promising and competitive for finding a local optimum point. In Part II, three formulas are designed by which a group of solutions are generated. This set of random formulas is hybridized with the globally convergent CG algorithm to obtain a hybrid stochastic conjugate gradient algorithm denoted by HSSZH. The HSSZH algorithm finds the approximate value of the global solution of a global optimization problem. Five combined stochastic conjugate gradient algorithms are constructed. The performance profiles are used to assess and compare the rendition of the family of hybrid stochastic conjugate gradient algorithms. The comparison results between our proposed HSSZH algorithm and four other hybrid stochastic conjugate gradient techniques demonstrate that the suggested HSSZH method is competitive with, and in all cases superior to, the four algorithms in terms of the efficiency, reliability and effectiveness to find the approximate solution of the global optimization problem that contains a non-convex function
State Transition Algorithm
In terms of the concepts of state and state transition, a new heuristic
random search algorithm named state transition algorithm is proposed. For
continuous function optimization problems, four special transformation
operators called rotation, translation, expansion and axesion are designed.
Adjusting measures of the transformations are mainly studied to keep the
balance of exploration and exploitation. Convergence analysis is also discussed
about the algorithm based on random search theory. In the meanwhile, to
strengthen the search ability in high dimensional space, communication strategy
is introduced into the basic algorithm and intermittent exchange is presented
to prevent premature convergence. Finally, experiments are carried out for the
algorithms. With 10 common benchmark unconstrained continuous functions used to
test the performance, the results show that state transition algorithms are
promising algorithms due to their good global search capability and convergence
property when compared with some popular algorithms.Comment: 18 pages, 28 figure
- …