310 research outputs found

    General Upper Bounds on the Runtime of Parallel Evolutionary Algorithms

    Get PDF
    We present a general method for analyzing the runtime of parallel evolutionary algorithms with spatially structured populations. Based on the fitness-level method, it yields upper bounds on the expected parallel runtime. This allows for a rigorous estimate of the speedup gained by parallelization. Tailored results are given for common migration topologies: ring graphs, torus graphs, hypercubes, and the complete graph. Example applications for pseudo-Boolean optimization show that our method is easy to apply and that it gives powerful results. In our examples the performance guarantees improve with the density of the topology. Surprisingly, even sparse topologies such as ring graphs lead to a significant speedup for many functions while not increasing the total number of function evaluations by more than a constant factor. We also identify which number of processors lead to the best guaranteed speedups, thus giving hints on how to parameterize parallel evolutionary algorithms

    How to analyse evolutionary algorithms

    Get PDF

    The SOS Platform: Designing, Tuning and Statistically Benchmarking Optimisation Algorithms

    Get PDF
    open access articleWe present Stochastic Optimisation Software (SOS), a Java platform facilitating the algorithmic design process and the evaluation of metaheuristic optimisation algorithms. SOS reduces the burden of coding miscellaneous methods for dealing with several bothersome and time-demanding tasks such as parameter tuning, implementation of comparison algorithms and testbed problems, collecting and processing data to display results, measuring algorithmic overhead, etc. SOS provides numerous off-the-shelf methods including: (1) customised implementations of statistical tests, such as the Wilcoxon rank-sum test and the Holm–Bonferroni procedure, for comparing the performances of optimisation algorithms and automatically generating result tables in PDF and formats; (2) the implementation of an original advanced statistical routine for accurately comparing couples of stochastic optimisation algorithms; (3) the implementation of a novel testbed suite for continuous optimisation, derived from the IEEE CEC 2014 benchmark, allowing for controlled activation of the rotation on each testbed function. Moreover, we briefly comment on the current state of the literature in stochastic optimisation and highlight similarities shared by modern metaheuristics inspired by nature. We argue that the vast majority of these algorithms are simply a reformulation of the same methods and that metaheuristics for optimisation should be simply treated as stochastic processes with less emphasis on the inspiring metaphor behind them

    Contributions to the mathematical modeling of estimation of distribution algorithms and pseudo-boolean functions

    Get PDF
    134 p.Maximice o minimice una función objetivo definida sobre un espacio discreto. Dado que la mayoría de dichos problemas no pueden ser resueltos mediante una búsqueda exhaustiva, su resolución se aproxima frecuentemente mediante algoritmos heurísticos. Sin embargo, no existe ningún algoritmo que se comporte mejor que el resto de algoritmos para resolver todas las instancias de cualquier problema. Por ello, el objetivo ideal es, dado una instancia de un problema, saber cuál es el algoritmo cuya resoluciones más eficiente. Las dos líneas principales de investigación para lograr dicho objetivo son estudiar las definiciones de los problemas y las posibles instancias que cada problema puede generar y el estudio delos diseños y características de los algoritmos. En esta tesis, se han tratado ambas lineas. Por un lado,hemos estudiado las funciones pseudo-Booleanas y varios problemas binarios específicos. Por otro lado,se ha presentado un modelado matemático para estudiar Algoritmos de Estimación de Distribuciones diseñados para resolver problemas basados en permutaciones. La principal motivación ha sido seguir progresando en este campo para comprender mejor las relaciones entre los Problemas de Optimización Combinatoria y los algoritmos de optimización

    04081 Abstracts Collection -- Theory of Evolutionary Algorithms

    Get PDF
    From 15.02.04 to 20.02.04, the Dagstuhl Seminar 04081 ``Theory of Evolutionary Algorithms\u27\u27 was held in the International Conference and Research Center (IBFI), Schloss Dagstuhl. During the seminar, several participants presented their current research, and ongoing work and open problems were discussed. Abstracts of the presentations given during the seminar as well as abstracts of seminar results and ideas are put together in this paper. The first section describes the seminar topics and goals in general. Links to extended abstracts or full papers are provided, if available

    Unifying a Geometric Framework of Evolutionary Algorithms and Elementary Landscapes Theory

    Get PDF
    Evolutionary algorithms (EAs) are randomised general-purpose strategies, inspired by natural evolution, often used for finding (near) optimal solutions to problems in combinatorial optimisation. Over the last 50 years, many theoretical approaches in evolutionary computation have been developed to analyse the performance of EAs, design EAs or measure problem difficulty via fitness landscape analysis. An open challenge is to formally explain why a general class of EAs perform better, or worse, than others on a class of combinatorial problems across representations. However, the lack of a general unified theory of EAs and fitness landscapes, across problems and representations, makes it harder to characterise pairs of general classes of EAs and combinatorial problems where good performance can be guaranteed provably. This thesis explores a unification between a geometric framework of EAs and elementary landscapes theory, not tied to a specific representation nor problem, with complementary strengths in the analysis of population-based EAs and combinatorial landscapes. This unification organises around three essential aspects: search space structure induced by crossovers, search behaviour of population-based EAs and structure of fitness landscapes. First, this thesis builds a crossover classification to systematically compare crossovers in the geometric framework and elementary landscapes theory, revealing a shared general subclass of crossovers: geometric recombination P-structures, which covers well-known crossovers. The crossover classification is then extended to a general framework for axiomatically analysing the population behaviour induced by crossover classes on associated EAs. This shows the shared general class of all EAs using geometric recombination P-structures, but no mutation, always do the same abstract form of convex evolutionary search. Finally, this thesis characterises a class of globally convex combinatorial landscapes shared by the geometric framework and elementary landscapes theory: abstract convex elementary landscapes. It is formally explained why geometric recombination P-structure EAs expectedly can outperform random search on abstract convex elementary landscapes related to low-order graph Laplacian eigenvalues. Altogether, this thesis paves a way towards a general unified theory of EAs and combinatorial fitness landscapes

    Noisy combinatorial optimisation with evolutionary algorithms

    Get PDF
    The determination of the efficient evolutionary optimisation approaches in solving noisy combinatorial problems is the main focus in this research. Initially, we present an empirical study of a range of evolutionary algorithms applied to various noisy combinatorial optimisation problems. There are four sets of experiments. The first looks at several toy problems, such as OneMax and other linear problems. We find that Univariate Marginal Distribution Algorithm (UMDA) and the Paired-Crossover Evolutionary Algorithm (PCEA) are the only ones able to cope robustly with noise, within a reasonable fixed time budget. In the second stage, UMDA and PCEA are then tested on more complex noisy problems: SubsetSum, Knapsack and SetCover. Both perform well under increasing levels of noise, with UMDA being the better of the two. In the third stage, we consider two noisy multi-objective problems (CountingOnesCountingZeros and a multi-objective formulation of SetCover). We compare several adaptations of UMDA for multi-objective problems with the Simple Evolutionary Multi-objective Optimiser (SEMO) and NSGA-II. In the last stage of empirical analysis, a realistic problem of the path planning for the ground surveillance with Unmanned Aerial Vehicles is considered. We conclude that UMDA, and its variants, can be highly effective on a variety of noisy combinatorial optimisation, outperforming many other evolutionary algorithms. Next, we study the use of voting mechanisms in populations, and introduce a new Voting algorithm which can solve OneMax and Jump in O(n log n), even for gaps as large as O(n). More significantly, the algorithm solves OneMax with added posterior noise in O(n log n), when the variance of the noise distribution is sigma2^2 = O(n) and in O(sigma2^2 log n) when the noise variance is greater than this. We assume only that the noise distribution has finite mean and variance and (for the larger noise case) that it is unimodal. Building upon this promising performance, we consider other noise models prevalent in optimisation and learning and show that the Voting algorithm has efficient performance in solving OneMax in presence of these noise variants. We also examine the performance on arbitrary linear and monotonic functions. The Voting algorithm fails on LeadingOnes but we give a variant which can solve the problem in O(n log n). We empirically study the use of voting in population based algorithms (UMDA, PCEA and cGA) and show that this can be effective for large population sizes
    • …
    corecore