1,495 research outputs found
A MOS-based Dynamic Memetic Differential Evolution Algorithm for Continuous Optimization: A Scalability Test
Continuous optimization is one of the areas with more activity in the field of heuristic optimization. Many algorithms have been proposed and compared on several benchmarks of functions, with different performance depending on the problems. For this reason, the combination of different search strategies seems desirable to obtain the best performance of each of these approaches. This contribution explores the use of a hybrid memetic algorithm based on the multiple offspring framework. The proposed algorithm combines the explorative/exploitative strength of two heuristic search methods that separately obtain very competitive results. This algorithm has been tested with the benchmark problems and conditions defined for the special issue of the Soft Computing Journal on Scalability of Evolutionary Algorithms and other Metaheuristics for Large Scale Continuous Optimization Problems. The proposed algorithm obtained the best results compared with both its composing algorithms and a set of reference algorithms that were proposed for the special issue
Memetic Artificial Bee Colony Algorithm for Large-Scale Global Optimization
Memetic computation (MC) has emerged recently as a new paradigm of efficient
algorithms for solving the hardest optimization problems. On the other hand,
artificial bees colony (ABC) algorithms demonstrate good performances when
solving continuous and combinatorial optimization problems. This study tries to
use these technologies under the same roof. As a result, a memetic ABC (MABC)
algorithm has been developed that is hybridized with two local search
heuristics: the Nelder-Mead algorithm (NMA) and the random walk with direction
exploitation (RWDE). The former is attended more towards exploration, while the
latter more towards exploitation of the search space. The stochastic adaptation
rule was employed in order to control the balancing between exploration and
exploitation. This MABC algorithm was applied to a Special suite on Large Scale
Continuous Global Optimization at the 2012 IEEE Congress on Evolutionary
Computation. The obtained results the MABC are comparable with the results of
DECC-G, DECC-G*, and MLCC.Comment: CONFERENCE: IEEE Congress on Evolutionary Computation, Brisbane,
Australia, 201
Enhancing Cooperative Coevolution for Large Scale Optimization by Adaptively Constructing Surrogate Models
It has been shown that cooperative coevolution (CC) can effectively deal with
large scale optimization problems (LSOPs) through a divide-and-conquer
strategy. However, its performance is severely restricted by the current
context-vector-based sub-solution evaluation method since this method needs to
access the original high dimensional simulation model when evaluating each
sub-solution and thus requires many computation resources. To alleviate this
issue, this study proposes an adaptive surrogate model assisted CC framework.
This framework adaptively constructs surrogate models for different
sub-problems by fully considering their characteristics. For the single
dimensional sub-problems obtained through decomposition, accurate enough
surrogate models can be obtained and used to find out the optimal solutions of
the corresponding sub-problems directly. As for the nonseparable sub-problems,
the surrogate models are employed to evaluate the corresponding sub-solutions,
and the original simulation model is only adopted to reevaluate some good
sub-solutions selected by surrogate models. By these means, the computation
cost could be greatly reduced without significantly sacrificing evaluation
quality. Empirical studies on IEEE CEC 2010 benchmark functions show that the
concrete algorithm based on this framework is able to find much better
solutions than the conventional CC algorithms and a non-CC algorithm even with
much fewer computation resources.Comment: arXiv admin note: text overlap with arXiv:1802.0974
Metaheuristic design of feedforward neural networks: a review of two decades of research
Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era
- …