533 research outputs found
Evolutionary Multiobjective Optimization Driven by Generative Adversarial Networks (GANs)
Recently, increasing works have proposed to drive evolutionary algorithms
using machine learning models. Usually, the performance of such model based
evolutionary algorithms is highly dependent on the training qualities of the
adopted models. Since it usually requires a certain amount of data (i.e. the
candidate solutions generated by the algorithms) for model training, the
performance deteriorates rapidly with the increase of the problem scales, due
to the curse of dimensionality. To address this issue, we propose a
multi-objective evolutionary algorithm driven by the generative adversarial
networks (GANs). At each generation of the proposed algorithm, the parent
solutions are first classified into real and fake samples to train the GANs;
then the offspring solutions are sampled by the trained GANs. Thanks to the
powerful generative ability of the GANs, our proposed algorithm is capable of
generating promising offspring solutions in high-dimensional decision space
with limited training data. The proposed algorithm is tested on 10 benchmark
problems with up to 200 decision variables. Experimental results on these test
problems demonstrate the effectiveness of the proposed algorithm
Gridless Evolutionary Approach for Line Spectral Estimation with Unknown Model Order
Gridless methods show great superiority in line spectral estimation. These
methods need to solve an atomic norm (i.e., the continuous analog of
norm) minimization problem to estimate frequencies and model order. Since
this problem is NP-hard to compute, relaxations of atomic norm, such as
nuclear norm and reweighted atomic norm, have been employed for promoting
sparsity. However, the relaxations give rise to a resolution limit,
subsequently leading to biased model order and convergence error. To overcome
the above shortcomings of relaxation, we propose a novel idea of simultaneously
estimating the frequencies and model order by means of the atomic norm.
To accomplish this idea, we build a multiobjective optimization model. The
measurment error and the atomic norm are taken as the two optimization
objectives. The proposed model directly exploits the model order via the atomic
norm, thus breaking the resolution limit. We further design a
variable-length evolutionary algorithm to solve the proposed model, which
includes two innovations. One is a variable-length coding and search strategy.
It flexibly codes and interactively searches diverse solutions with different
model orders. These solutions act as steppingstones that help fully exploring
the variable and open-ended frequency search space and provide extensive
potentials towards the optima. Another innovation is a model order pruning
mechanism, which heuristically prunes less contributive frequencies within the
solutions, thus significantly enhancing convergence and diversity. Simulation
results confirm the superiority of our approach in both frequency estimation
and model order selection.Comment: This work has been submitted to the IEEE for possible publication.
Copyright may be transferred without notice, after which this version may no
longer be accessibl
Cultural Algorithm based on Decomposition to solve Optimization Problems
Decomposition is used to solve optimization problems by introducing many simple scalar optimization subproblems and optimizing them simultaneously. Dynamic Multi-Objective Optimization Problems (DMOP) have several objective functions and constraints that vary over time. As a consequence of such dynamic changes, the optimal solutions may vary over time, affecting the performance of convergence. In this thesis, we propose a new Cultural Algorithm (CA) based on decomposition (CA/D). The objective of the CA/D algorithm is to decompose DMOP into a number of subproblems that can be optimized using the information shared by neighboring problems. The proposed CA/D approach is evaluated using a number of CEC 2015 optimization benchmark functions. When compared to CA, Multi-population CA (MPCA), and MPCA incorporating game strategies (MPCA-GS), the results obtained showed that CA/D outperformed them in 7 out of the 15 benchmark functions
Automated Design of Metaheuristic Algorithms: A Survey
Metaheuristics have gained great success in academia and practice because
their search logic can be applied to any problem with available solution
representation, solution quality evaluation, and certain notions of locality.
Manually designing metaheuristic algorithms for solving a target problem is
criticized for being laborious, error-prone, and requiring intensive
specialized knowledge. This gives rise to increasing interest in automated
design of metaheuristic algorithms. With computing power to fully explore
potential design choices, the automated design could reach and even surpass
human-level design and could make high-performance algorithms accessible to a
much wider range of researchers and practitioners. This paper presents a broad
picture of automated design of metaheuristic algorithms, by conducting a survey
on the common grounds and representative techniques in terms of design space,
design strategies, performance evaluation strategies, and target problems in
this field
Multiobjective Design and Optimization of Polymer Flood Performance
The multiobjective genetic algorithm can be used to optimize two conflicting objectives, oil production and polymer utility factor in polymer flood design. This approach provides a set of optimal solutions which can be considered as trade-off curve (Pareto front) to maximize oil production while preserving polymer performance. Then an optimal polymer flood design can be considered from post-optimization analysis. A 2D synthetic example, and a 3D field-scale application, accounting for geologic uncertainty, showed that beyond the optimal design, a relatively minor increase in oil production requires much more polymer injection and the polymer utility factor increases substantially
Metaheuristic design of feedforward neural networks: a review of two decades of research
Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era
- …