250 research outputs found

    The SOS Platform: Designing, Tuning and Statistically Benchmarking Optimisation Algorithms

    Get PDF
    open access articleWe present Stochastic Optimisation Software (SOS), a Java platform facilitating the algorithmic design process and the evaluation of metaheuristic optimisation algorithms. SOS reduces the burden of coding miscellaneous methods for dealing with several bothersome and time-demanding tasks such as parameter tuning, implementation of comparison algorithms and testbed problems, collecting and processing data to display results, measuring algorithmic overhead, etc. SOS provides numerous off-the-shelf methods including: (1) customised implementations of statistical tests, such as the Wilcoxon rank-sum test and the Holm–Bonferroni procedure, for comparing the performances of optimisation algorithms and automatically generating result tables in PDF and formats; (2) the implementation of an original advanced statistical routine for accurately comparing couples of stochastic optimisation algorithms; (3) the implementation of a novel testbed suite for continuous optimisation, derived from the IEEE CEC 2014 benchmark, allowing for controlled activation of the rotation on each testbed function. Moreover, we briefly comment on the current state of the literature in stochastic optimisation and highlight similarities shared by modern metaheuristics inspired by nature. We argue that the vast majority of these algorithms are simply a reformulation of the same methods and that metaheuristics for optimisation should be simply treated as stochastic processes with less emphasis on the inspiring metaphor behind them

    Novel Memetic Computing Structures for Continuous Optimisation

    Get PDF
    This thesis studies a class of optimisation algorithms, namely Memetic Computing Structures, and proposes a novel set of promising algorithms that move the first step towards an implementation for the automatic generation of optimisation algorithms for continuous domains. This thesis after a thorough review of local search algorithms and popular meta-heuristics, focuses on Memetic Computing in terms of algorithm structures and design philosophy. In particular, most of the design carried out during my doctoral studies is inspired by the lex parsimoniae, aka Ockham’s Razor. It has been shown how simple algorithms, when well implemented can outperform complex implementations. In order to achieve this aim, the design is always carried out by attempting to identify the role of each algorithmic component/operator. In this thesis, on the basis of this logic, a set of variants of a recently proposed algorithms are presented. Subsequently a novel memetic structure, namely Parallel Memetic Structure is proposed and tested against modern algorithms representing the state of the art in optimisation. Furthermore, an initial prototype of an automatic design platform is also included. This prototype performs an analysis on separability of the optimisation problem and, on the basis of the analysis results, designs some parts of the parallel structure. Promising results are included. Finally, an investigation of the correlation among the variables and problem dimensionality has been performed. An extremely interesting finding of this thesis work is that the degree of correlation among the variables decreases when the dimensionality increases. As a direct consequence of this fact, large scale problems are to some extent easier to handle than problems in low dimensionality since, due to the lack of correlation among the variables, they can effectively be tackled by an algorithm that performs moves along the axes

    A self-adaptive multimeme memetic algorithm co-evolving utility scores to control genetic operators and their parameter settings

    Get PDF
    Memetic algorithms are a class of well-studied metaheuristics which combine evolutionary algorithms and local search techniques. A meme represents contagious piece of information in an adaptive information sharing system. The canonical memetic algorithm uses a fixed meme, denoting a hill climbing operator, to improve each solution in a population during the evolutionary search process. Given global parameters and multiple parametrised operators, adaptation often becomes a crucial constituent in the design of MAs. In this study, a self-adaptive self-configuring steady-state multimeme memetic algorithm (SSMMA) variant is proposed. Along with the individuals (solutions), SSMMA co-evolves memes, encoding the utility score for each algorithmic component choice and relevant parameter setting option. An individual uses tournament selection to decide which operator and parameter setting to employ at a given step. The performance of the proposed algorithm is evaluated on six combinatorial optimisation problems from a cross-domain heuristic search benchmark. The results indicate the success of SSMMA when compared to the static Mas as well as widely used self-adaptive Multimeme Memetic Algorithm from the scientific literature

    HyperSPAM: A study on hyper-heuristic coordination strategies in the continuous domain

    Get PDF
    The file attached to this record is the author's final peer reviewed version. The Publisher's final version can be found by following the DOI link.This article proposes a simplistic algorithmic framework, namely hyperSPAM, composed of three search algorithms for addressing continuous optimisation problems. The Covariance Matrix Adaptation Evolution Strategy (CMAES) is activated at the beginning of the optimisation process as a preprocessing component for a limited budget. Subsequently, the produced solution is fed to the other two single-solution search algorithms. The first performs moves along the axes while the second makes use of a matrix orthogonalization to perform diagonal moves. Four coordination strategies, in the fashion of hyperheuristics, have been used to coordinate the two single-solution algorithms. One of them is a simple randomized criterion while the other three are based on a success based reward mechanism. The four implementations of the hyperSPAM framework have been tested and compared against each other and modern metaheuristics on an extensive set of problems including theoretical functions and real-world engineering problems. Numerical results show that the different versions of the framework display broadly a similar performance. One of the reward schemes appears to be marginally better than the others. The simplistic random coordination also displays a very good performance. All the implementations of hyperSPAM significantly outperform the other algorithms used for comparison

    The SOS Platform: Designing, Tuning and Statistically Benchmarking Optimisation Algorithms

    Get PDF
    We present Stochastic Optimisation Software (SOS), a Java platform facilitating the algorithmic design process and the evaluation of metaheuristic optimisation algorithms. SOS reduces the burden of coding miscellaneous methods for dealing with several bothersome and time-demanding tasks such as parameter tuning, implementation of comparison algorithms and testbed problems, collecting and processing data to display results, measuring algorithmic overhead, etc. SOS provides numerous off-the-shelf methods including: (1) customised implementations of statistical tests, such as the Wilcoxon rank-sum test and the Holm–Bonferroni procedure, for comparing the performances of optimisation algorithms and automatically generating result tables in PDF and LATEX formats; (2) the implementation of an original advanced statistical routine for accurately comparing couples of stochastic optimisation algorithms; (3) the implementation of a novel testbed suite for continuous optimisation, derived from the IEEE CEC 2014 benchmark, allowing for controlled activation of the rotation on each testbed function. Moreover, we briefly comment on the current state of the literature in stochastic optimisation and highlight similarities shared by modern metaheuristics inspired by nature. We argue that the vast majority of these algorithms are simply a reformulation of the same methods and that metaheuristics for optimisation should be simply treated as stochastic processes with less emphasis on the inspiring metaphor behind them

    Cooperative co-evolution with differential grouping for large scale optimization

    Get PDF
    Cooperative co-evolution has been introduced into evolutionary algorithms with the aim of solving increasingly complex optimization problems through a divide-and-conquer paradigm. In theory, the idea of co-adapted subcomponents is desirable for solving large-scale optimization problems. However, in practice, without prior knowledge about the problem, it is not clear how the problem should be decomposed. In this paper, we propose an automatic decomposition strategy called differential grouping that can uncover the underlying interaction structure of the decision variables and form subcomponents such that the interdependence between them is kept to a minimum. We show mathematically how such a decomposition strategy can be derived from a definition of partial separability. The empirical studies show that such near-optimal decomposition can greatly improve the solution quality on large-scale global optimization problems. Finally, we show how such an automated decomposition allows for a better approximation of the contribution of various subcomponents, leading to a more efficient assignment of the computational budget to various subcomponents

    A multi-cycled sequential memetic computing approach for constrained optimisation

    Get PDF
    In this paper, we propose a multi-cycled sequential memetic computing structure for constrained optimisation. The structure is composed of multiple evolutionary cycles. At each cycle, an evolutionary algorithm is considered as an operator, and connects with a local optimiser. This structure enables the learning of useful knowledge from previous cycles and the transfer of the knowledge to facilitate search in latter cycles. Specifically, we propose to apply an estimation of distribution algorithm (EDA) to explore the search space until convergence at each cycle. A local optimiser, called DONLP2, is then applied to improve the best solution found by the EDA. New cycle starts after the local improvement if the computation budget has not been exceeded. In the developed EDA, an adaptive fully-factorized multivariate probability model is proposed. A learning mechanism, implemented as the guided mutation operator, is adopted to learn useful knowledge from previous cycles. The developed algorithm was experimentally studied on the benchmark problems in the CEC 2006 and 2010 competition. Experimental studies have shown that the developed probability model exhibits excellent exploration capability and the learning mechanism can significantly improve the search efficiency under certain conditions. The comparison against some well-known algorithms showed the superiority of the developed algorithm in terms of the consumed fitness evaluations and the solution quality
    • …
    corecore