2,356 research outputs found

    A memetic particle swarm optimisation algorithm for dynamic multi-modal optimisation problems

    Get PDF
    Copyright @ 2011 Taylor & Francis.Many real-world optimisation problems are both dynamic and multi-modal, which require an optimisation algorithm not only to find as many optima under a specific environment as possible, but also to track their moving trajectory over dynamic environments. To address this requirement, this article investigates a memetic computing approach based on particle swarm optimisation for dynamic multi-modal optimisation problems (DMMOPs). Within the framework of the proposed algorithm, a new speciation method is employed to locate and track multiple peaks and an adaptive local search method is also hybridised to accelerate the exploitation of species generated by the speciation method. In addition, a memory-based re-initialisation scheme is introduced into the proposed algorithm in order to further enhance its performance in dynamic multi-modal environments. Based on the moving peaks benchmark problems, experiments are carried out to investigate the performance of the proposed algorithm in comparison with several state-of-the-art algorithms taken from the literature. The experimental results show the efficiency of the proposed algorithm for DMMOPs.This work was supported by the Key Program of National Natural Science Foundation (NNSF) of China under Grant no. 70931001, the Funds for Creative Research Groups of China under Grant no. 71021061, the National Natural Science Foundation (NNSF) of China under Grant 71001018, Grant no. 61004121 and Grant no. 70801012 and the Fundamental Research Funds for the Central Universities Grant no. N090404020, the Engineering and Physical Sciences Research Council (EPSRC) of UK under Grant no. EP/E060722/01 and Grant EP/E060722/02, and the Hong Kong Polytechnic University under Grant G-YH60

    Three variants of three Stage Optimal Memetic Exploration for handling non-separable fitness landscapes

    Get PDF
    The file attached to this record is the authors final peer reviewed version. The publisher's final version can be found by following the DOI link.Three Stage Optimal Memetic Exploration (3SOME) is a recently proposed algorithmic framework which sequentially perturbs a single solution by means of three operators. Although 3SOME proved to be extremely successful at handling high-dimensional multi-modal landscapes, its application to non-separable fitness functions present some flaws. This paper proposes three possible variants of the original 3SOME algorithm aimed at improving its performance on non-separable problems. The first variant replaces one of the 3SOME operators, namely the middle distance exploration, with a rotation-invariant Differential Evolution (DE) mutation scheme, which is applied on three solutions sampled in a progressively shrinking search space. In the second proposed mechanism, a micro-population rotation-invariant DE is integrated within the algorithmic framework. The third approach employs the search logic (1+1)-Covariance Matrix Adaptation Evolution Strategy, aka (1+1)-CMA-ES. In the latter scheme, a Covariance Matrix adapts to the landscape during the optimization in order to determine the most promising search directions. Numerical results show that, at the cost of a higher complexity, the three approaches proposed are able to improve upon 3SOME performance for non-separable problems without an excessive performance deterioration in the other problems

    A Differential Evolution Framework with Ensemble of Parameters and Strategies and Pool of Local Search Algorithms

    Get PDF
    The file attached to this record is the author's final peer reviewed version. The publisher's final version can be found by following the DOI link.The ensemble structure is a computational intelligence supervised strategy consisting of a pool of multiple operators that compete among each other for being selected, and an adaptation mechanism that tends to reward the most successful operators. In this paper we extend the idea of the ensemble to multiple local search logics. In a memetic fashion, the search structure of an ensemble framework cooperatively/competitively optimizes the problem jointly with a pool of diverse local search algorithms. In this way, the algorithm progressively adapts to a given problem and selects those search logics that appear to be the most appropriate to quickly detect high quality solutions. The resulting algorithm, namely Ensemble of Parameters and Strategies Differential Evolution empowered by Local Search (EPSDE-LS), is evaluated on multiple testbeds and dimensionality values. Numerical results show that the proposed EPSDE-LS robustly displays a very good performance in comparison with some of the state-of-the-art algorithms

    Compact Optimization Algorithms with Re-sampled Inheritance

    Get PDF
    The file attached to this record is the author's final peer reviewed version.Compact optimization algorithms are a class of Estimation of Distribution Algorithms (EDAs) characterized by extremely limited memory requirements (hence they are called \compact"). As all EDAs, compact algorithms build and update a probabilistic model of the distribution of solutions within the search space, as opposed to population-based algorithms that instead make use of an explicit population of solutions. In addition to that, to keep their memory consumption low, compact algorithms purposely employ simple probabilistic models that can be described with a small number of parameters. Despite their simplicity, compact algorithms have shown good performances on a broad range of benchmark functions and real-world problems. However, compact algorithms also come with some drawbacks, i.e. they tend to premature convergence and show poorer performance on non-separable problems. To overcome these limitations, here we investigate a possible algorithmic scheme obtained by combining compact algorithms with a non-disruptive restart mechanism taken from the literature, named Re-Sampled Inheritance (RI). The resulting compact algorithms with RI are tested on the CEC 2014 benchmark functions. The numerical results show on the one hand that the use of RI consistently enhances the performances of compact algorithms, still keeping a limited usage of memory. On the other hand, our experiments show that among the tested algorithms, the best performance is obtained by compact Differential Evolution with RI

    Compact real-valued teaching-learning based optimization with the applications to neural network training

    Get PDF
    The majority of embedded systems are designed for specific applications, often associated with limited hardware resources in order to meet various and sometime conflicting requirements such as cost, speed, size and performance. Advanced intelligent heuristic optimization algorithms have been widely used in solving engineering problems. However, they might not be applicable to embedded systems, which often have extremely limited memory size. In this paper, a new compact teaching-learning based optimization method for solving global continuous problems is proposed, particularly aiming for neural network training in portable artificial intelligent (AI) devices. Comprehensive numerical experiments on benchmark problems and the training of two popular neural network systems verify that the new compact algorithm is capable of maintaining the high performance while the memory requirement is significantly reduced. It offers a promising tool for continuous optimization problems including the training of neural networks for intelligent embedded systems with limited memory resources
    • ā€¦
    corecore