1,022 research outputs found
Recommended from our members
Combinatorial optimization and metaheuristics
Today, combinatorial optimization is one of the youngest and most active areas of discrete mathematics. It is a branch of optimization in applied mathematics and computer science, related to operational research, algorithm theory and computational complexity theory. It sits at the intersection of several fields, including artificial intelligence, mathematics and software engineering. Its increasing interest arises for the fact that a large number of scientific and industrial problems can be formulated as abstract combinatorial optimization problems, through graphs and/or (integer) linear programs. Some of these problems have polynomial-time (“efficient”) algorithms, while most of them are NP-hard, i.e. it is not proved that they can be solved in polynomial-time. Mainly, it means that it is not possible to guarantee that an exact solution to the problem can be found and one has to settle for an approximate solution with known performance guarantees. Indeed, the goal of approximate methods is to find “quickly” (reasonable run-times), with “high” probability, provable “good” solutions (low error from the real optimal solution). In the last 20 years, a new kind of algorithm commonly called metaheuristics have emerged in this class, which basically try to combine heuristics in high level frameworks aimed at efficiently and effectively exploring the search space. This report briefly outlines the components, concepts, advantages and disadvantages of different metaheuristic approaches from a conceptual point of view, in order to analyze their similarities and differences. The two very significant forces of intensification and diversification, that mainly determine the behavior of a metaheuristic, will be pointed out. The report concludes by exploring the importance of hybridization and integration methods
Adaptive primal-dual genetic algorithms in dynamic environments
This article is placed here with permission of IEEE - Copyright @ 2010 IEEERecently, there has been an increasing interest in applying genetic algorithms (GAs) in dynamic environments. Inspired by the complementary and dominance mechanisms in nature, a primal-dual GA (PDGA) has been proposed for dynamic optimization problems (DOPs). In this paper, an important operator in PDGA, i.e., the primal-dual mapping (PDM) scheme, is further investigated to improve the robustness and adaptability of PDGA in dynamic environments. In the improved scheme, two different probability-based PDM operators, where the mapping probability of each allele in the chromosome string is calculated through the statistical information of the distribution of alleles in the corresponding gene locus over the population, are effectively combined according to an adaptive Lamarckian learning mechanism. In addition, an adaptive dominant replacement scheme, which can probabilistically accept inferior chromosomes, is also introduced into the proposed algorithm to enhance the diversity level of the population. Experimental results on a series of dynamic problems generated from several stationary benchmark problems show that the proposed algorithm is a good optimizer for DOPs.This work was supported in part by the National Nature Science Foundation of China (NSFC) under Grant 70431003 and Grant
70671020, by the National Innovation Research Community Science Foundation
of China under Grant 60521003, by the National Support Plan of China under Grant 2006BAH02A09, by the Engineering and Physical Sciences
Research Council (EPSRC) of U.K. under Grant EP/E060722/1, and by the
Hong Kong Polytechnic University Research Grants under Grant G-YH60
Neuroevolution for solving multiobjective knapsack problems
The multiobjective knapsack problem (MOKP) is an important combinatorial problem that arises in various applications, including resource allocation, computer science and finance. When tackling this problem by evolutionary multiobjective optimization algorithms (EMOAs), it has been demonstrated that traditional recombination operators acting on binary solution representations are susceptible to a loss of diversity and poor scalability. To address those issues, we propose to use artificial neural networks for generating solutions by performing a binary classification of items using the information about their profits and weights. As gradient-based learning cannot be used when target values are unknown, neuroevolution is adapted to adjust the neural network parameters. The main contribution of this study resides in developing a solution encoding and genotype-phenotype mapping for EMOAs to solve MOKPs. The proposal is implemented within a state-of-the-art EMOA and benchmarked against traditional variation operators based on binary crossovers. The obtained experimental results indicate a superior performance of the proposed approach. Furthermore, it is advantageous in terms of scalability and can be readily incorporated into different EMOAs.Portuguese “Fundação para a Ciência e Tecnologia” under grant PEst-C/CTM/LA0025/2013 (Projecto Estratégico - LA 25 - 2013-2014 - Strategic Project - LA 25 - 2013-2014
Improving Time and Memory Efficiency of Genetic Algorithms by Storing Populations as Minimum Spanning Trees of Patches
In many applications of evolutionary algorithms the computational cost of
applying operators and storing populations is comparable to the cost of fitness
evaluation. Furthermore, by knowing what exactly has changed in an individual
by an operator, it is possible to recompute fitness value much more efficiently
than from scratch. The associated time and memory improvements have been
available for simple evolutionary algorithms, few specific genetic algorithms
and in the context of gray-box optimization, but not for all algorithms, and
the main reason is that it is difficult to achieve in algorithms using large
arbitrarily structured populations.
This paper makes a first step towards improving this situation. We show that
storing the population as a minimum spanning tree, where vertices correspond to
individuals but only contain meta-information about them, and edges store
structural differences, or patches, between the individuals, is a viable
alternative to the straightforward implementation. Our experiments suggest that
significant, even asymptotic, improvements -- including execution of crossover
operators! -- can be achieved in terms of both memory usage and computational
costs.Comment: Accepted to the GECCO'23 conference, EvoSoft worksho
Explicit Building-Block Multiobjective Genetic Algorithms: Theory, Analysis, and Developing
This dissertation research emphasizes explicit Building Block (BB) based MO EAs performance and detailed symbolic representation. An explicit BB-based MOEA for solving constrained and real-world MOPs is developed the Multiobjective Messy Genetic Algorithm II (MOMGA-II) which is designed to validate symbolic BB concepts. The MOMGA-II demonstrates that explicit BB-based MOEAs provide insight into solving difficult MOPs that is generally not realized through the use of implicit BB-based MOEA approaches. This insight is necessary to increase the effectiveness of all MOEA approaches. In order to increase MOEA computational efficiency parallelization of MOEAs is addressed. Communications between processors in a parallel MOEA implementation is extremely important, hence innovative migration and replacement schemes for use in parallel MOEAs are detailed and tested. These parallel concepts support the development of the first explicit BB-based parallel MOEA the pMOMGA-II. MOEA theory is also advanced through the derivation of the first MOEA population sizing theory. The multiobjective population sizing theory presented derives the MOEA population size necessary in order to achieve good results within a specified level of confidence. Just as in the single objective approach the MOEA population sizing theory presents a very conservative sizing estimate. Validated results illustrate insight into building block phenomena good efficiency excellent effectiveness and motivation for future research in the area of explicit BB-based MOEAs. Thus the generic results of this research effort have applicability that aid in solving many different MOPs
Multi-objective genetic programming with partial sampling and its extension to many-objective
This paper describes a technique on an optimization of tree-structure data by of multi-objective evolutionary algorithm, or multi-objective genetic programming. GP induces bloat of the tree structure as one of the major problem. The cause of bloat is that the tree structure obtained by the crossover operator grows bigger and bigger but its evaluation does not improve. To avoid the risk of bloat, a partial sampling operator is proposed as a mating operator. The size of the tree and a structural distance are introduced into the measure of the tree-structure data as the objective functions in addition to the index of the goodness of tree structure. GP is defined as a three-objective optimization problem. SD is also applied for the ranking of parent individuals instead to the crowding distance of the conventional NSGA-II. When the index of the goodness of tree-structure data is two or more, the number of objective functions in the above problem becomes four or more. We also propose an effective many-objective EA applicable to such the many-objective GP. We focus on NSGA-II based on Pareto partial dominance (NSGA-II-PPD). NSGA-II-PPD requires beforehand a combination list of the number of objective functions to be used for Pareto partial dominance (PPD). The contents of the combination list greatly influence the optimization result. We propose to schedule a parameter r meaning the subset size of objective functions for PPD and to eliminate individuals created by the mating having the same contents as the individual of the archive set
- …