21 research outputs found

    Do not Choose Representation just Change: An Experimental Study in States based EA

    Get PDF
    Our aim in this paper is to analyse the phenotypic effects (evolvability) of diverse coding conversion operators in an instance of the states based evolutionary algorithm (SEA). Since the representation of solutions or the selection of the best encoding during the optimization process has been proved to be very important for the efficiency of evolutionary algorithms (EAs), we will discuss a strategy of coupling more than one representation and different procedures of conversion from one coding to another during the search. Elsewhere, some EAs try to use multiple representations (SM-GA, SEA, etc.) in intention to benefit from the characteristics of each of them. In spite of those results, this paper shows that the change of the representation is also a crucial approach to take into consideration while attempting to increase the performances of such EAs. As a demonstrative example, we use a two states SEA (2-SEA) which has two identical search spaces but different coding conversion operators. The results show that the way of changing from one coding to another and not only the choice of the best representation nor the representation itself is very advantageous and must be taken into account in order to well-desing and improve EAs execution

    On the Movement of Vertex Fixed Points in the Simple GA

    Get PDF
    ABSTRACT The Vose dynamical system model of the simple genetic algorithm models the behavior of this algorithm for large population sizes and is the basis of the exact Markov chain model. Populations consisting of multiple copies of one individual correspond to vertices of the simplex. For zero mutation, these are fixed points of the dynamical system and absorbing states of the Markov chain. For proportional selection, the asymptotic stability of vertex fixed points is understood from previous work. We derive the eigenvalues of the differential at vertex fixed points of the dynamical system model for tournament selection. We show that as mutation increases from zero, hyperbolic asymptotically stable fixed points move into the simplex, and hyperbolic asymptotically unstable fixed points move outside of the simplex. We calculate the derivative of local path of the fixed point with respect to the mutation rate for proportional selection. Simulation analysis shows how fixed points bifurcate with larger changes in the mutation rate and changes in the crossover rate

    A probabilistic cooperative-competitive hierarchical search model.

    Get PDF
    by Wong Yin Bun, Terence.Thesis (M.Phil.)--Chinese University of Hong Kong, 1998.Includes bibliographical references (leaves 99-104).Abstract also in Chinese.List of Figures --- p.ixList of Tables --- p.xiChapter I --- Preliminary --- p.1Chapter 1 --- Introduction --- p.2Chapter 1.1 --- Thesis themes --- p.4Chapter 1.1.1 --- Dynamical view of landscape --- p.4Chapter 1.1.2 --- Bottom-up self-feedback algorithm with memory --- p.4Chapter 1.1.3 --- Cooperation and competition --- p.5Chapter 1.1.4 --- Contributions to genetic algorithms --- p.5Chapter 1.2 --- Thesis outline --- p.5Chapter 1.3 --- Contribution at a glance --- p.6Chapter 1.3.1 --- Problem --- p.6Chapter 1.3.2 --- Approach --- p.7Chapter 1.3.3 --- Contributions --- p.7Chapter 2 --- Background --- p.8Chapter 2.1 --- Iterative stochastic searching algorithms --- p.8Chapter 2.1.1 --- The algorithm --- p.8Chapter 2.1.2 --- Stochasticity --- p.10Chapter 2.2 --- Fitness landscapes and its relation to neighborhood --- p.12Chapter 2.2.1 --- Direct searching --- p.12Chapter 2.2.2 --- Exploration and exploitation --- p.12Chapter 2.2.3 --- Fitness landscapes --- p.13Chapter 2.2.4 --- Neighborhood --- p.16Chapter 2.3 --- Species formation methods --- p.17Chapter 2.3.1 --- Crowding methods --- p.17Chapter 2.3.2 --- Deterministic crowding --- p.18Chapter 2.3.3 --- Sharing method --- p.18Chapter 2.3.4 --- Dynamic niching --- p.19Chapter 2.4 --- Summary --- p.21Chapter II --- Probabilistic Binary Hierarchical Search --- p.22Chapter 3 --- The basic algorithm --- p.23Chapter 3.1 --- Introduction --- p.23Chapter 3.2 --- Search space reduction with binary hierarchy --- p.25Chapter 3.3 --- Search space modeling --- p.26Chapter 3.4 --- The information processing cycle --- p.29Chapter 3.4.1 --- Local searching agents --- p.29Chapter 3.4.2 --- Global environment --- p.30Chapter 3.4.3 --- Cooperative refinement and feedback --- p.33Chapter 3.5 --- Enhancement features --- p.34Chapter 3.5.1 --- Fitness scaling --- p.34Chapter 3.5.2 --- Elitism --- p.35Chapter 3.6 --- Illustration of the algorithm behavior --- p.36Chapter 3.6.1 --- Test problem --- p.36Chapter 3.6.2 --- Performance study --- p.38Chapter 3.6.3 --- Benchmark tests --- p.45Chapter 3.7 --- Discussion and analysis --- p.45Chapter 3.7.1 --- Hierarchy of partitions --- p.45Chapter 3.7.2 --- Availability of global information --- p.47Chapter 3.7.3 --- Adaptation --- p.47Chapter 3.8 --- Summary --- p.48Chapter III --- Cooperation and Competition --- p.50Chapter 4 --- High-dimensionality --- p.51Chapter 4.1 --- Introduction --- p.51Chapter 4.1.1 --- The challenge of high-dimensionality --- p.51Chapter 4.1.2 --- Cooperation - A solution to high-dimensionality --- p.52Chapter 4.2 --- Probabilistic Cooperative Binary Hierarchical Search --- p.52Chapter 4.2.1 --- Decoupling --- p.52Chapter 4.2.2 --- Cooperative fitness --- p.53Chapter 4.2.3 --- The cooperative model --- p.54Chapter 4.3 --- Empirical performance study --- p.56Chapter 4.3.1 --- pBHS versus pcBHS --- p.56Chapter 4.3.2 --- Scaling behavior of pcBHS --- p.60Chapter 4.3.3 --- Benchmark test --- p.62Chapter 4.4 --- Summary --- p.63Chapter 5 --- Deception --- p.65Chapter 5.1 --- Introduction --- p.65Chapter 5.1.1 --- The challenge of deceptiveness --- p.65Chapter 5.1.2 --- Competition: A solution to deception --- p.67Chapter 5.2 --- Probabilistic cooperative-competitive binary hierarchical search --- p.67Chapter 5.2.1 --- Overview --- p.68Chapter 5.2.2 --- The cooperative-competitive model --- p.68Chapter 5.3 --- Empirical performance study --- p.70Chapter 5.3.1 --- Goldberg's deceptive function --- p.70Chapter 5.3.2 --- "Shekel family - S5, S7, and S10" --- p.73Chapter 5.4 --- Summary --- p.74Chapter IV --- Finale --- p.78Chapter 6 --- A new genetic operator --- p.79Chapter 6.1 --- Introduction --- p.79Chapter 6.2 --- Variants of the integration --- p.80Chapter 6.2.1 --- Fixed-fraction-of-all --- p.83Chapter 6.2.2 --- Fixed-fraction-of-best --- p.83Chapter 6.2.3 --- Best-from-both --- p.84Chapter 6.3 --- Empricial performance study --- p.84Chapter 6.4 --- Summary --- p.88Chapter 7 --- Conclusion and Future work --- p.89Chapter A --- The pBHS Algorithm --- p.91Chapter A.1 --- Overview --- p.91Chapter A.2 --- Details --- p.91Chapter B --- Test problems --- p.96Bibliography --- p.9

    A Computational View on Natural Evolution: On the Rigorous Analysis of the Speed of Adaptation

    Get PDF
    Inspired by Darwin’s ideas, Turing (1948) proposed an evolutionary search as an automated problem solving approach. Mimicking natural evolution, evolutionary algorithms evolve a set of solutions through the repeated application of the evolutionary operators (mutation, recombination and selection). Evolutionary algorithms belong to the family of black box algorithms which are general purpose optimisation tools. They are typically used when no good specific algorithm is known for the problem at hand and they have been reported to be surprisingly effective (Eiben and Smith, 2015; Sarker et al., 2002). Interestingly, although evolutionary algorithms are heavily inspired by natural evolution, their study has deviated from the study of evolution by the population genetics community. We believe that this is a missed opportunity and that both fields can benefit from an interdisciplinary collaboration. The question of how long it takes for a natural population to evolve complex adaptations has fascinated researchers for decades. We will argue that this is an equivalent research question to the runtime analysis of algorithms. By making use of the methods and techniques used in both fields, we will derive plenty of meaningful results for both communities, proving that this interdisciplinary approach is effective and relevant. We will apply the tools used in the theoretical analysis of evolutionary algorithms to quantify the complexity of adaptive walks on many landscapes, illustrating how the structure of the fitness landscape and the parameter conditions can impose limits to adaptation. Furthermore, as geneticists use diffusion theory to track the change in the allele frequencies of a population, we will develop a brand new model to analyse the dynamics of evolutionary algorithms. Our model, based on stochastic differential equations, will allow to describe not only the expected behaviour, but also to measure how much the process might deviate from that expectation

    An Analysis of Particle Swarm Optimizers

    Get PDF
    Many scientific, engineering and economic problems involve the optimisation of a set of parameters. These problems include examples like minimising the losses in a power grid by finding the optimal configuration of the components, or training a neural network to recognise images of people's faces. Numerous optimisation algorithms have been proposed to solve these problems, with varying degrees of success. The Particle Swarm Optimiser (PSO) is a relatively new technique that has been empirically shown to perform well on many of these optimisation problems. This thesis presents a theoretical model that can be used to describe the long-term behaviour of the algorithm. An enhanced version of the Particle Swarm Optimiser is constructed and shown to have guaranteed convergence on local minima. This algorithm is extended further, resulting in an algorithm with guaranteed convergence on global minima. A model for constructing cooperative PSO algorithms is developed, resulting in the introduction of two new PSO-based algorithms. Empirical results are presented to support the theoretical properties predicted by the various models, using synthetic benchmark functions to investigate specific properties. The various PSO-based algorithms are then applied to the task of training neural networks, corroborating the results obtained on the synthetic benchmark functions.Thesis (PhD)--University of Pretoria, 2007.Computer ScienceUnrestricte

    Complexity Theory for Discrete Black-Box Optimization Heuristics

    Full text link
    A predominant topic in the theory of evolutionary algorithms and, more generally, theory of randomized black-box optimization techniques is running time analysis. Running time analysis aims at understanding the performance of a given heuristic on a given problem by bounding the number of function evaluations that are needed by the heuristic to identify a solution of a desired quality. As in general algorithms theory, this running time perspective is most useful when it is complemented by a meaningful complexity theory that studies the limits of algorithmic solutions. In the context of discrete black-box optimization, several black-box complexity models have been developed to analyze the best possible performance that a black-box optimization algorithm can achieve on a given problem. The models differ in the classes of algorithms to which these lower bounds apply. This way, black-box complexity contributes to a better understanding of how certain algorithmic choices (such as the amount of memory used by a heuristic, its selective pressure, or properties of the strategies that it uses to create new solution candidates) influences performance. In this chapter we review the different black-box complexity models that have been proposed in the literature, survey the bounds that have been obtained for these models, and discuss how the interplay of running time analysis and black-box complexity can inspire new algorithmic solutions to well-researched problems in evolutionary computation. We also discuss in this chapter several interesting open questions for future work.Comment: This survey article is to appear (in a slightly modified form) in the book "Theory of Randomized Search Heuristics in Discrete Search Spaces", which will be published by Springer in 2018. The book is edited by Benjamin Doerr and Frank Neumann. Missing numbers of pointers to other chapters of this book will be added as soon as possibl
    corecore