547 research outputs found

    Challenging test problems for multi- and many-objective optimization

    Get PDF
    In spite of the extensive studies that have been conducted regarding the construction of multi-objective test problems, researchers have mainly focused their interests on designing complicated search spaces, disregarding, in many cases, the design of the Pareto optimal fronts of the problems. In this regard, the work related to scalable multi-objective test problems – i.e., problems that can be formulated for an arbitrary number of objectives – has been much less studied. This paper introduces a new set of continuous and box-constrained multi-objective test problems which are scalable in both the number of objectives and in the number of decision variables. Each test problem included in the proposed test suite has a peculiar Pareto front different from those observed in the existing scalable multi-objective test suites. In addition to different Pareto fronts, the proposed test suite introduces features related to the search space that place obstacles that complicate exploring Pareto optimal solutions. Such features can be easily switched on and off by the user to analyze specific mechanisms of multi-objective evolutionary algorithms (MOEAs). The components used in the proposed test suite can be used as a toolkit to construct new test instances not included in this set of problems. To illustrate the use and difficulties of the proposed test suite, some experiments are presented adopting three MOEAs using selection mechanisms based on Pareto optimality, decomposition, and a performance indicator (hypervolume)

    A Faster Algorithm for Calculating Hypervolume

    Get PDF
    We present an algorithm for calculating hypervolume exactly, the Hypervolume by Slicing Objectives (HSO) algorithm, that is faster than any that has previously been published. HSO processes objectives instead of points, an idea that has been considered before but that has never been properly evaluated in the literature. We show that both previously studied exact hypervolume algorithms are exponential in at least the number of objectives and that although HSO is also exponential in the number of objectives in the worst case, it runs in significantly less time, i.e., two to three orders of magnitude less for randomly generated and benchmark data in three to eight objectives. Thus, HSO increases the utility of hypervolume, both as a metric for general optimization algorithms and as a diversity mechanism for evolutionary algorithm

    Decomposition-Based-Sorting and Angle-Based-Selection for Evolutionary Multiobjective and Many-Objective Optimization

    Get PDF
    Multiobjective evolutionary algorithm based on decomposition (MOEA/D) decomposes a multiobjective optimization problem (MOP) into a number of scalar optimization subproblems and then solves them in parallel. In many MOEA/D variants, each subproblem is associated with one and only one solution. An underlying assumption is that each subproblem has a different Pareto-optimal solution, which may not be held, for irregular Pareto fronts (PFs), e.g., disconnected and degenerate ones. In this paper, we propose a new variant of MOEA/D with sorting-and-selection (MOEA/D-SAS). Different from other selection schemes, the balance between convergence and diversity is achieved by two distinctive components, decomposition-based-sorting (DBS) and angle-based-selection (ABS). DBS only sorts L{L} closest solutions to each subproblem to control the convergence and reduce the computational cost. The parameter L{L} has been made adaptive based on the evolutionary process. ABS takes use of angle information between solutions in the objective space to maintain a more fine-grained diversity. In MOEA/D-SAS, different solutions can be associated with the same subproblems; and some subproblems are allowed to have no associated solution, more flexible to MOPs or many-objective optimization problems (MaOPs) with different shapes of PFs. Comprehensive experimental studies have shown that MOEA/D-SAS outperforms other approaches; and is especially effective on MOPs or MaOPs with irregular PFs. Moreover, the computational efficiency of DBS and the effects of ABS in MOEA/D-SAS are also investigated and discussed in detail

    On the utilization of pair-potential energy functions in multi-objective optimization

    Get PDF
    In evolutionary multi-objective optimization (EMO), the pair-potential energy functions (PPFs) have been used to construct diversity-preserving mechanisms to improve Pareto front approximations. Despite PPFs have shown promising results when dealing with different Pareto front geometries, there are still some open research questions to improve the way we employ them. In this paper, we answer three important questions: (1) what is the effect of a crucial parameter of some PPFs?, (2) how do we set the optimal parameter value?, and (3) what is the best PPF in EMO? To solve these questions, we designed a brand-new fast algorithm to generate an approximate solution to a PPF-based subset selection problem and, then, we conducted a comprehensive parametrical study to predict the optimal parameter values using a deep neural network. To show the effectiveness of the PPF-based diversity-preserving mechanisms, we selected two application cases: the generation of reference point sets of benchmark problems (DTLZ, WFG, IDTLZ, IWFG, IMOP, and Viennet) with different Pareto front shapes, and the definition of a PPF-based archive that can be coupled to any multi-objective evolutionary algorithm to construct well-diversified Pareto front approximations. Using several diversity indicators, it is shown that the utilization of PPF-based mechanisms lead to good Pareto front approximations regardless of the Pareto front shape

    Quality Measures of Parameter Tuning for Aggregated Multi-Objective Temporal Planning

    Get PDF
    Parameter tuning is recognized today as a crucial ingredient when tackling an optimization problem. Several meta-optimization methods have been proposed to find the best parameter set for a given optimization algorithm and (set of) problem instances. When the objective of the optimization is some scalar quality of the solution given by the target algorithm, this quality is also used as the basis for the quality of parameter sets. But in the case of multi-objective optimization by aggregation, the set of solutions is given by several single-objective runs with different weights on the objectives, and it turns out that the hypervolume of the final population of each single-objective run might be a better indicator of the global performance of the aggregation method than the best fitness in its population. This paper discusses this issue on a case study in multi-objective temporal planning using the evolutionary planner DaE-YAHSP and the meta-optimizer ParamILS. The results clearly show how ParamILS makes a difference between both approaches, and demonstrate that indeed, in this context, using the hypervolume indicator as ParamILS target is the best choice. Other issues pertaining to parameter tuning in the proposed context are also discussed.Comment: arXiv admin note: substantial text overlap with arXiv:1305.116
    • …
    corecore