3,002 research outputs found
On-line Search History-assisted Restart Strategy for Covariance Matrix Adaptation Evolution Strategy
Restart strategy helps the covariance matrix adaptation evolution strategy
(CMA-ES) to increase the probability of finding the global optimum in
optimization, while a single run CMA-ES is easy to be trapped in local optima.
In this paper, the continuous non-revisiting genetic algorithm (cNrGA) is used
to help CMA-ES to achieve multiple restarts from different sub-regions of the
search space. The CMA-ES with on-line search history-assisted restart strategy
(HR-CMA-ES) is proposed. The entire on-line search history of cNrGA is stored
in a binary space partitioning (BSP) tree, which is effective for performing
local search. The frequently sampled sub-region is reflected by a deep position
in the BSP tree. When leaf nodes are located deeper than a threshold, the
corresponding sub-region is considered a region of interest (ROI). In
HR-CMA-ES, cNrGA is responsible for global exploration and suggesting ROI for
CMA-ES to perform an exploitation within or around the ROI. CMA-ES restarts
independently in each suggested ROI. The non-revisiting mechanism of cNrGA
avoids to suggest the same ROI for a second time. Experimental results on the
CEC 2013 and 2017 benchmark suites show that HR-CMA-ES performs better than
both CMA-ES and cNrGA. A positive synergy is observed by the memetic
cooperation of the two algorithms.Comment: 8 pages, 9 figure
Continuous non-revisiting genetic algorithm
The non-revisiting genetic algorithm (NrGA) is extended to handle continuous search space. The extended NrGA model, Continuous NrGA (cNrGA), employs the same tree-structure archive of NrGA to memorize the evaluated solutions, in which the search space is divided into non-overlapped partitions according to the distribution of the solutions. cNrGA is a bi-modulus evolutionary algorithm consisting of the genetic algorithm module (GAM) and the adaptive mutation module (AMM). When GAM generates an offspring, the offspring is sent to AMM and is mutated according to the density of the solutions stored in the memory archive. For a point in the search space with high solution-density, it infers a high probability that the point is close to the optimum and hence a near search is suggested. Alternatively, a far search is recommended for a point with low solution-density. Benefitting from the space partitioning scheme, a fast solution-density approximation is obtained. Also, the adaptive mutation scheme naturally avoid the generation of out-of-bound solutions. The performance of cNrGA is tested on 14 benchmark functions on dimensions ranging from 2 to 40. It is compared with real coded GA, differential evolution, covariance matrix adaptation evolution strategy and two improved particle swarm optimization. The simulation results show that cNrGA outperforms the other algorithms for multi-modal function optimization.published_or_final_versio
Continuous non-revisiting genetic algorithm with random search space re-partitioning and one-gene-flip mutation
Special Session on Evolutionary Computer VisionIn continuous non-revisiting genetic algorithm (cNrGA), the solution set with different order leads to different density estimation and hence different mutation step size. As a result, the performance of cNrGA depends on the order of the evaluated solutions. In this paper, we propose to remove this dependence by a search space re-partitioning strategy. At each iteration, the strategy re-shuffles the solutions into random order. The re-ordered sequence is then used to construct a new density tree, which leads to a new space partition sets. Afterwards, instead of randomly picking a mutant within a partition, a new adaptive one-gene-flip mutation is applied. Motivated from the fact that the proposed adaptive mutation concerns only small amount of partitions, we propose a new density tree construction algorithm. This algorithm refuses to partition the sub-regions which do not contain any individual to be mutated, which simplifies the tree topology as well as speeds up the construction time. The new cNrGA integrated with the proposed re-partitioning strategy (cNrGA/RP/OGF) is examined on 19 benchmark functions at dimensions ranging from 2 to 40. The simulation results show that cNrGA/RP/OGF is significantly superior to the original cNrGA at most of the test functions. Its average performance is also better than those of six benchmark EAs. © 2010 IEEE.published_or_final_versio
A study of operator and parameter choices in non-revisiting genetic algorithm
We study empirically the effects of operator and parameter choices on the performance of the non-revisiting genetic algorithm (NrGA). For a suite of 14 benchmark functions that include both uni-modal and multi-modal functions, it is found that NrGA is insensitive to the axis resolution of the problem, which is a good feature. From the empirical experiments, for operators, it is found that crossover is an essential operator for NrGA, and the best crossover operator is uniform crossover, while the best selection operator is elitist selection. For parameters, a small population, with a population size strictly larger than 1, should be used; the performance is monotonically increasing with crossover rate and the optimal crossover rate is 0.5. The results of this paper provide empirical guidelines for operator designs and parameter settings of NrGA. © 2009 IEEE.published_or_final_versio
Continuous non-revisiting genetic algorithm with overlapped search sub-region
In continuous non-revisiting genetic algorithm (cNrGA), search space is partitioned into sub-regions according to the distribution of evaluated solutions. The partitioned subregion serves as mutation range such that the corresponding mutation is adaptive and parameter-less. As pointed out by Chow and Yuen, the boundary condition of the mutation in cNrGA is too restricted that the exploitative power of cNrGA is reduced. In this paper, we tackle this structural problem of cNrGA by a new formulation of mutation range. When sub-region is formulated as which certain overlap exists between adjacent sub-regions, this creates a soft boundary and it allows individual move from a sub-region to another with better fitness. This modified cNrGA is named cNrGA with overlapped search sub-region (cNrGA/OL/OGF). By comparing with another work on this problem, Continuous non-revisiting genetic algorithm with randomly re-partitioned BSP tree (cNrGA/RP/OGF), it has an advantage on processing speed. The proposed algorithm is examined on 34 benchmark functions at dimensions ranging from 2 to 40. The results show that the proposed algorithm is superior to the original cNrGA, cNrGA/RP/OGF and covariance matrix adaptation evolutionary strategy (CMA-ES). © 2012 IEEE.published_or_final_versio
A non-revisiting simulated annealing algorithm
In this article, a non-revisiting simulated annealing algorithm (NrSA) is proposed. NrSA is an integration of the non-revisiting scheme and standard simulated annealing (SA). It guarantees that every generated neighbor must not be visited before. This property leads to reduction on the computation cost on evaluating time consuming and expensive objective functions such as surface registration, optimized design and energy management of heating, ventilating and air conditioning systems. Meanwhile, the prevention on function re-evaluation also speeds up the convergence. Furthermore, due to the nature of the non-revisiting scheme, the returned non-revisited solutions from the scheme can be treated as self-adaptive solutions, such that no parametric neighbor picking scheme is involved in NrSA. Thus NrSA can be identified as a parameter-less SA. The simulation results show that NrSA is superior to adaptive SA (ASA) on both uni-modal and multi-modal functions with dimension up to 40. We also illustrate that the overhead and archive size of NrSA are insignificant, so it is practical for real world applications. © 2008 IEEE.published_or_final_versio
Recommended from our members
Illuminating meaningful diversity in complex feature spaces through adaptive grid-based genetic algorithms
In many fields there exist problems for which multiple solutions of suitably high performance may be found across distinct regions of the search space. Optimisation of the search towards including these distinct solutions is important not only to understanding these spaces but also to avoiding local optima. This is the goal of a type of genetic algorithms called illumination algorithms. In Chapter 2, we demonstrate the use of an illumination algorithm in the exploration of networks sharing only a given set of structural features (valid networks). This method produces a population of valid networks that are more diverse than those produced using state of the art methods, however, it was found to be too inefficient to be usable in real-world problems. Additionally, setting an appropriate resolution of the search requires some amount of prior knowledge of the space of solutions. Addressing this problem is the focus of Chapter 3, in which we develop three extensions to the method: a) an exact method of mutation whereby only valid networks are explored, b) an adaptive mechanism for setting the resolution of the search, c) a principle for tuning mutations parameters to the search’ s resolution. We show that with these additions our method is able to increase the diversity of solutions found in significantly fewer iterations. Finally, in Chapter 4 we expand our method for use in more general problem spaces. We benchmark it against the state of the art. In all tested landscapes, we show that our method is able to identify more meaningful niches in the spaces in the same number of iterations. We conclude by highlighting the limits of our framework and discuss further directions
Recommended from our members
Combinatorial optimization and metaheuristics
Today, combinatorial optimization is one of the youngest and most active areas of discrete mathematics. It is a branch of optimization in applied mathematics and computer science, related to operational research, algorithm theory and computational complexity theory. It sits at the intersection of several fields, including artificial intelligence, mathematics and software engineering. Its increasing interest arises for the fact that a large number of scientific and industrial problems can be formulated as abstract combinatorial optimization problems, through graphs and/or (integer) linear programs. Some of these problems have polynomial-time (“efficient”) algorithms, while most of them are NP-hard, i.e. it is not proved that they can be solved in polynomial-time. Mainly, it means that it is not possible to guarantee that an exact solution to the problem can be found and one has to settle for an approximate solution with known performance guarantees. Indeed, the goal of approximate methods is to find “quickly” (reasonable run-times), with “high” probability, provable “good” solutions (low error from the real optimal solution). In the last 20 years, a new kind of algorithm commonly called metaheuristics have emerged in this class, which basically try to combine heuristics in high level frameworks aimed at efficiently and effectively exploring the search space. This report briefly outlines the components, concepts, advantages and disadvantages of different metaheuristic approaches from a conceptual point of view, in order to analyze their similarities and differences. The two very significant forces of intensification and diversification, that mainly determine the behavior of a metaheuristic, will be pointed out. The report concludes by exploring the importance of hybridization and integration methods
- …