154 research outputs found

    Evolutionary Multiobjective Optimization Driven by Generative Adversarial Networks (GANs)

    Get PDF
    Recently, increasing works have proposed to drive evolutionary algorithms using machine learning models. Usually, the performance of such model based evolutionary algorithms is highly dependent on the training qualities of the adopted models. Since it usually requires a certain amount of data (i.e. the candidate solutions generated by the algorithms) for model training, the performance deteriorates rapidly with the increase of the problem scales, due to the curse of dimensionality. To address this issue, we propose a multi-objective evolutionary algorithm driven by the generative adversarial networks (GANs). At each generation of the proposed algorithm, the parent solutions are first classified into real and fake samples to train the GANs; then the offspring solutions are sampled by the trained GANs. Thanks to the powerful generative ability of the GANs, our proposed algorithm is capable of generating promising offspring solutions in high-dimensional decision space with limited training data. The proposed algorithm is tested on 10 benchmark problems with up to 200 decision variables. Experimental results on these test problems demonstrate the effectiveness of the proposed algorithm

    Hybrid non-dominated sorting genetic algorithm with adaptive operators selection

    Get PDF
    Multiobjective optimization entails minimizing or maximizing multiple objective functions subject to a set of constraints. Many real world applications can be formulated as multi-objective optimization problems (MOPs), which often involve multiple conflicting objectives to be optimized simultaneously. Recently, a number of multi-objective evolutionary algorithms (MOEAs) were developed suggested for these MOPs as they do not require problem specific information. They find a set of non-dominated solutions in a single run. The evolutionary process on which they are based, typically relies on a single genetic operator. Here, we suggest an algorithm which uses a basket of search operators. This is because it is never easy to choose the most suitable operator for a given problem. The novel hybrid non-dominated sorting genetic algorithm (HNSGA) introduced here in this paper and tested on the ZDT (Zitzler-Deb-Thiele) and CEC’09 (2009 IEEE Conference on Evolutionary Computations) benchmark problems specifically formulated for MOEAs. Numerical results prove that the proposed algorithm is competitive with state-of-the-art MOEAs

    Enhanced Version of Multi-algorithm Genetically Adaptive for Multiobjective optimization

    Get PDF
    Abstract: Multi-objective EAs (MOEAs) are well established population-based techniques for solving various search and optimization problems. MOEAs employ different evolutionary operators to evolve populations of solutions for approximating the set of optimal solutions of the problem at hand in a single simulation run. Different evolutionary operators suite different problems. The use of multiple operators with a self-adaptive capability can further improve the performance of existing MOEAs. This paper suggests an enhanced version of a genetically adaptive multi-algorithm for multi-objective (AMAL-GAM) optimisation which includes differential evolution (DE), particle swarm optimization (PSO), simulated binary crossover (SBX), Pareto archive evolution strategy (PAES) and simplex crossover (SPX) for population evolution during the course of optimization. We examine the performance of this enhanced version of AMALGAM experimentally over two different test suites, the ZDT test problems and the test instances designed recently for the special session on MOEA?s competition at the Congress of Evolutionary Computing of 2009 (CEC?09). The suggested algorithm has found better approximate solutions on most test problems in terms of inverted generational distance (IGD) as the metric indicator. - See more at: http://thesai.org/Publications/ViewPaper?Volume=6&Issue=12&Code=ijacsa&SerialNo=37#sthash.lxkuyzEf.dpu

    An adaptive evolutionary multi-objective approach based on simulated annealing

    Get PDF
    A multi-objective optimization problem can be solved by decomposing it into one or more single objective subproblems in some multi-objective metaheuristic algorithms. Each subproblem corresponds to one weighted aggregation function. For example, MOEA/D is an evolutionary multi-objective optimization (EMO) algorithm that attempts to optimize multiple subproblems simultaneously by evolving a population of solutions. However, the performance of MOEA/D highly depends on the initial setting and diversity of the weight vectors. In this paper, we present an improved version of MOEA/D, called EMOSA, which incorporates an advanced local search technique (simulated annealing) and adapts the search directions (weight vectors) corresponding to various subproblems. In EMOSA, the weight vector of each subproblem is adaptively modified at the lowest temperature in order to diversify the search towards the unexplored parts of the Pareto-optimal front. Our computational results show that EMOSA outperforms six other well-established multi-objective metaheuristic algorithms on both the (constrained)multi-objective knapsack problemand the (unconstrained) multi-objective traveling salesman problem. Moreover, the effects of the main algorithmic components and parameter sensitivities on the search performance of EMOSA are experimentally investigated

    Hybrid adaptive evolutionary algorithm based on decomposition

    Get PDF
    The performance of search operators varies across the different stages of the search/optimization process of evolutionary algorithms (EAs). In general, a single search operator may not do well in all these stages when dealing with different optimization and search problems. To mitigate this, adaptive search operator schemes have been introduced. The idea is that when a search operator hits a difficult patch (under-performs) in the search space, the EA scheme “reacts” to that by potentially calling upon a different search operator. Hence, several multiple-search operator schemes have been proposed and employed within EA. In this paper, a hybrid adaptive evolutionary algorithm based on decomposition (HAEA/D) that employs four different crossover operators is suggested. Its performance has been evaluated on the well-known IEEE CEC’09 test instances. HAEA/D has generated promising results which compare well against several well-known algorithms including MOEA/D, on a number of metrics such as the inverted generational distance (IGD), the hyper-volume, the Gamma and Delta functions. These results are included and discussed in this paper

    A survey on handling computationally expensive multiobjective optimization problems with evolutionary algorithms

    Get PDF
    This is the author accepted manuscript. The final version is available from Springer Verlag via the DOI in this record.Evolutionary algorithms are widely used for solving multiobjective optimization problems but are often criticized because of a large number of function evaluations needed. Approximations, especially function approximations, also referred to as surrogates or metamodels are commonly used in the literature to reduce the computation time. This paper presents a survey of 45 different recent algorithms proposed in the literature between 2008 and 2016 to handle computationally expensive multiobjective optimization problems. Several algorithms are discussed based on what kind of an approximation such as problem, function or fitness approximation they use. Most emphasis is given to function approximation-based algorithms. We also compare these algorithms based on different criteria such as metamodeling technique and evolutionary algorithm used, type and dimensions of the problem solved, handling constraints, training time and the type of evolution control. Furthermore, we identify and discuss some promising elements and major issues among algorithms in the literature related to using an approximation and numerical settings used. In addition, we discuss selecting an algorithm to solve a given computationally expensive multiobjective optimization problem based on the dimensions in both objective and decision spaces and the computation budget available.The research of Tinkle Chugh was funded by the COMAS Doctoral Program (at the University of Jyväskylä) and FiDiPro Project DeCoMo (funded by Tekes, the Finnish Funding Agency for Innovation), and the research of Dr. Karthik Sindhya was funded by SIMPRO project funded by Tekes as well as DeCoMo

    Parallel Multi-Objective Evolutionary Algorithms: A Comprehensive Survey

    Get PDF
    Multi-Objective Evolutionary Algorithms (MOEAs) are powerful search techniques that have been extensively used to solve difficult problems in a wide variety of disciplines. However, they can be very demanding in terms of computational resources. Parallel implementations of MOEAs (pMOEAs) provide considerable gains regarding performance and scalability and, therefore, their relevance in tackling computationally expensive applications. This paper presents a survey of pMOEAs, describing a refined taxonomy, an up-to-date review of methods and the key contributions to the field. Furthermore, some of the open questions that require further research are also briefly discussed

    On the Effect of the Cooperation of Indicator-Based Multiobjective Evolutionary Algorithms

    Get PDF
    For almost 20 years, quality indicators (QIs) have promoted the design of new selection mechanisms of multiobjective evolutionary algorithms (MOEAs). Each indicator-based MOEA (IB-MOEA) has specific search preferences related to its baseline QI, producing Pareto front approximations with different properties. In consequence, an IB-MOEA based on a single QI has a limited scope of multiobjective optimization problems (MOPs) in which it is expected to have a good performance. This issue is emphasized when the associated Pareto front geometries are highly irregular. In order to overcome these issues, we propose here an island-based multiindicator algorithm (IMIA) that takes advantage of the search biases of multiple IB-MOEAs through a cooperative scheme. Our experimental results show that the cooperation of multiple IB-MOEAs allows IMIA to perform more robustly (considering several QIs) than the panmictic versions of its baseline IB-MOEAs as well as several state-of-the-art MOEAs. Additionally, IMIA shows a Pareto-front-shape invariance property, which makes it a remarkable optimizer when tackling MOPs with complex Pareto front geometries
    • …
    corecore