137,516 research outputs found
Ensemble Learning for Free with Evolutionary Algorithms ?
Evolutionary Learning proceeds by evolving a population of classifiers, from
which it generally returns (with some notable exceptions) the single
best-of-run classifier as final result. In the meanwhile, Ensemble Learning,
one of the most efficient approaches in supervised Machine Learning for the
last decade, proceeds by building a population of diverse classifiers. Ensemble
Learning with Evolutionary Computation thus receives increasing attention. The
Evolutionary Ensemble Learning (EEL) approach presented in this paper features
two contributions. First, a new fitness function, inspired by co-evolution and
enforcing the classifier diversity, is presented. Further, a new selection
criterion based on the classification margin is proposed. This criterion is
used to extract the classifier ensemble from the final population only
(Off-line) or incrementally along evolution (On-line). Experiments on a set of
benchmark problems show that Off-line outperforms single-hypothesis
evolutionary learning and state-of-art Boosting and generates smaller
classifier ensembles
Assessing the robustness of parsimonious predictions for gene neighborhoods from reconciled phylogenies
The availability of a large number of assembled genomes opens the way to
study the evolution of syntenic character within a phylogenetic context. The
DeCo algorithm, recently introduced by B{\'e}rard et al. allows the computation
of parsimonious evolutionary scenarios for gene adjacencies, from pairs of
reconciled gene trees. Following the approach pioneered by Sturmfels and
Pachter, we describe how to modify the DeCo dynamic programming algorithm to
identify classes of cost schemes that generates similar parsimonious
evolutionary scenarios for gene adjacencies, as well as the robustness to
changes to the cost scheme of evolutionary events of the presence or absence of
specific ancestral gene adjacencies. We apply our method to six thousands
mammalian gene families, and show that computing the robustness to changes to
cost schemes provides new and interesting insights on the evolution of gene
adjacencies and the DeCo model.Comment: Accepted, to appear in ISBRA - 11th International Symposium on
Bioinformatics Research and Applications - 2015, Jun 2015, Norfolk, Virginia,
United State
Neuroevolution on the Edge of Chaos
Echo state networks represent a special type of recurrent neural networks.
Recent papers stated that the echo state networks maximize their computational
performance on the transition between order and chaos, the so-called edge of
chaos. This work confirms this statement in a comprehensive set of experiments.
Furthermore, the echo state networks are compared to networks evolved via
neuroevolution. The evolved networks outperform the echo state networks,
however, the evolution consumes significant computational resources. It is
demonstrated that echo state networks with local connections combine the best
of both worlds, the simplicity of random echo state networks and the
performance of evolved networks. Finally, it is shown that evolution tends to
stay close to the ordered side of the edge of chaos.Comment: To appear in Proceedings of the Genetic and Evolutionary Computation
Conference 2017 (GECCO '17
Distributed evolutionary algorithms and their models: A survey of the state-of-the-art
The increasing complexity of real-world optimization problems raises new challenges to evolutionary computation. Responding to these challenges, distributed evolutionary computation has received considerable attention over the past decade. This article provides a comprehensive survey of the state-of-the-art distributed evolutionary algorithms and models, which have been classified into two groups according to their task division mechanism. Population-distributed models are presented with master-slave, island, cellular, hierarchical, and pool architectures, which parallelize an evolution task at population, individual, or operation levels. Dimension-distributed models include coevolution and multi-agent models, which focus on dimension reduction. Insights into the models, such as synchronization, homogeneity, communication, topology, speedup, advantages and disadvantages are also presented and discussed. The study of these models helps guide future development of different and/or improved algorithms. Also highlighted are recent hotspots in this area, including the cloud and MapReduce-based implementations, GPU and CUDA-based implementations, distributed evolutionary multiobjective optimization, and real-world applications. Further, a number of future research directions have been discussed, with a conclusion that the development of distributed evolutionary computation will continue to flourish
- …