8,763 research outputs found
Towards the Evolution of Novel Vertical-Axis Wind Turbines
Renewable and sustainable energy is one of the most important challenges
currently facing mankind. Wind has made an increasing contribution to the
world's energy supply mix, but still remains a long way from reaching its full
potential. In this paper, we investigate the use of artificial evolution to
design vertical-axis wind turbine prototypes that are physically instantiated
and evaluated under approximated wind tunnel conditions. An artificial neural
network is used as a surrogate model to assist learning and found to reduce the
number of fabrications required to reach a higher aerodynamic efficiency,
resulting in an important cost reduction. Unlike in other approaches, such as
computational fluid dynamics simulations, no mathematical formulations are used
and no model assumptions are made.Comment: 14 pages, 11 figure
Evolutionary Multiobjective Optimization Driven by Generative Adversarial Networks (GANs)
Recently, increasing works have proposed to drive evolutionary algorithms
using machine learning models. Usually, the performance of such model based
evolutionary algorithms is highly dependent on the training qualities of the
adopted models. Since it usually requires a certain amount of data (i.e. the
candidate solutions generated by the algorithms) for model training, the
performance deteriorates rapidly with the increase of the problem scales, due
to the curse of dimensionality. To address this issue, we propose a
multi-objective evolutionary algorithm driven by the generative adversarial
networks (GANs). At each generation of the proposed algorithm, the parent
solutions are first classified into real and fake samples to train the GANs;
then the offspring solutions are sampled by the trained GANs. Thanks to the
powerful generative ability of the GANs, our proposed algorithm is capable of
generating promising offspring solutions in high-dimensional decision space
with limited training data. The proposed algorithm is tested on 10 benchmark
problems with up to 200 decision variables. Experimental results on these test
problems demonstrate the effectiveness of the proposed algorithm
Learning Opposites with Evolving Rules
The idea of opposition-based learning was introduced 10 years ago. Since then
a noteworthy group of researchers has used some notions of oppositeness to
improve existing optimization and learning algorithms. Among others,
evolutionary algorithms, reinforcement agents, and neural networks have been
reportedly extended into their opposition-based version to become faster and/or
more accurate. However, most works still use a simple notion of opposites,
namely linear (or type- I) opposition, that for each assigns its
opposite as . This, of course, is a very naive estimate of
the actual or true (non-linear) opposite , which has been
called type-II opposite in literature. In absence of any knowledge about a
function that we need to approximate, there seems to be no
alternative to the naivety of type-I opposition if one intents to utilize
oppositional concepts. But the question is if we can receive some level of
accuracy increase and time savings by using the naive opposite estimate
according to all reports in literature, what would we be able to
gain, in terms of even higher accuracies and more reduction in computational
complexity, if we would generate and employ true opposites? This work
introduces an approach to approximate type-II opposites using evolving fuzzy
rules when we first perform opposition mining. We show with multiple examples
that learning true opposites is possible when we mine the opposites from the
training data to subsequently approximate .Comment: Accepted for publication in The 2015 IEEE International Conference on
Fuzzy Systems (FUZZ-IEEE 2015), August 2-5, 2015, Istanbul, Turke
Freeze-drying modeling and monitoring using a new neuro-evolutive technique
This paper is focused on the design of a black-box model for the process of freeze-drying of pharmaceuticals. A new methodology based on a self-adaptive differential evolution scheme is combined with a back-propagation algorithm, as local search method, for the simultaneous structural and parametric optimization of the model represented by a neural network. Using the model of the freeze-drying process, both the temperature and the residual ice content in the product vs. time can be determine off-line, given the values of the operating conditions (the temperature of the heating shelf and the pressure in the drying chamber). This makes possible to understand if the maximum temperature allowed by the product is trespassed and when the sublimation drying is complete, thus providing a valuable tool for recipe design and optimization. Besides, the black box model can be applied to monitor the freeze-drying process: in this case, the measurement of product temperature is used as input variable of the neural network in order to provide in-line estimation of the state of the product (temperature and residual amount of ice). Various examples are presented and discussed, thus pointing out the strength of the too
A Comparison of Nature Inspired Algorithms for Multi-threshold Image Segmentation
In the field of image analysis, segmentation is one of the most important
preprocessing steps. One way to achieve segmentation is by mean of threshold
selection, where each pixel that belongs to a determined class islabeled
according to the selected threshold, giving as a result pixel groups that share
visual characteristics in the image. Several methods have been proposed in
order to solve threshold selectionproblems; in this work, it is used the method
based on the mixture of Gaussian functions to approximate the 1D histogram of a
gray level image and whose parameters are calculated using three nature
inspired algorithms (Particle Swarm Optimization, Artificial Bee Colony
Optimization and Differential Evolution). Each Gaussian function approximates
thehistogram, representing a pixel class and therefore a threshold point.
Experimental results are shown, comparing in quantitative and qualitative
fashion as well as the main advantages and drawbacks of each algorithm, applied
to multi-threshold problem.Comment: 16 pages, this is a draft of the final version of the article sent to
the Journa
Metaheuristic design of feedforward neural networks: a review of two decades of research
Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era
- …