156 research outputs found

    Genetic Drift in Genetic Algorithm Selection Schemes

    No full text
    A method for calculating genetic drift in terms of changing population fitness variance is presented. The method allows for an easy comparison of different selection schemes and exact analytical results are derived for traditional generational selection, steady-state selection with varying generation gap, a simple model of Eshelman's CHC algorithm, and evolution strategies. The effects of changing genetic drift on the convergence of a GA are demonstrated empirically

    Phase Transitions and Symmetry Breaking in Genetic Algorithms with Crossover

    No full text
    In this paper, we consider the role of the crossover operator in genetic algorithms. Specifically, we study optimisation problems that exhibit many local optima and consider how crossover affects the rate at which the population breaks the symmetry of the problem. As an example of such a problem, we consider the subset sum problem. In so doing, we demonstrate a previously unobserved phenomenon, whereby the genetic algorithm with crossover exhibits a critical mutation rate, at which its performance sharply diverges from that of the genetic algorithm without crossover. At this critical mutation rate, the genetic algorithm with crossover exhibits a rapid increase in population diversity. We calculate the details of this phenomenon on a simple instance of the subset sum problem and show that it is a classic phase transition between ordered and disordered populations. Finally, we show that this critical mutation rate corresponds to the transition between the genetic algorithm accelerating or preventing symmetry breaking and that the critical mutation rate represents an optimum in terms of the balance of exploration and exploitation within the algorithm

    Learning short synfire chains by self-organization

    No full text
    We develop a model of cortical coding of stimuli by the sequences of activation patterns that they ignite in an initially random network. Hebbian learning then stabilizes these sequences, making them attractors of the dynamics. There is a competition between the capacity of the network and the stability of the sequences; for small stability parameter epsilon (the strength of the mean stabilizing PSP in the neurons in a learned sequence) the capacity is proportional to 1/epsilon2. For epsilon of the order of or less than the PSPs of the untrained network, the capacity exceeds that for sequences learned from tabula rasa

    Annealing schedule from population dynamics

    Full text link
    We introduce a dynamical annealing schedule for population-based optimization algorithms with mutation. On the basis of a statistical mechanics formulation of the population dynamics, the mutation rate adapts to a value maximizing expected rewards at each time step. Thereby, the mutation rate is eliminated as a free parameter from the algorithm.Comment: 6 pages RevTeX, 4 figures PostScript; to be published in Phys. Rev.

    Learning short synfire chains by self-organization

    Full text link
    corecore