7 research outputs found

    Analyzing Adaptive Parameter Landscapes in Parameter Adaptation Methods for Differential Evolution

    Full text link
    Since the scale factor and the crossover rate significantly influence the performance of differential evolution (DE), parameter adaptation methods (PAMs) for the two parameters have been well studied in the DE community. Although PAMs can sufficiently improve the effectiveness of DE, PAMs are poorly understood (e.g., the working principle of PAMs). One of the difficulties in understanding PAMs comes from the unclarity of the parameter space that consists of the scale factor and the crossover rate. This paper addresses this issue by analyzing adaptive parameter landscapes in PAMs for DE. First, we propose a concept of an adaptive parameter landscape, which captures a moment in a parameter adaptation process. For each iteration, each individual in the population has its adaptive parameter landscape. Second, we propose a method of analyzing adaptive parameter landscapes using a 1-step-lookahead greedy improvement metric. Third, we examine adaptive parameter landscapes in PAMs by using the proposed method. Results provide insightful information about PAMs in DE.Comment: This is an accepted version of a paper published in the proceedings of GECCO 202

    Analysis of Evolutionary Algorithms Using Multi-objective Parameter Tuning

    No full text
    Evolutionary Algorithms (EAs) and other metaheuristics are greatly affected by the choice of their parameters, not only as regards the precision of the solutions found, but also for repeatability, robustness, speed of convergence, and other properties. Most of these performance criteria are often conflicting with one another. In our work, we see the problem of EAs' parameter selection and tuning as a multi-objective optimization problem, in which the criteria to be optimized are precision and speed of convergence. We propose EMOPaT (Evolutionary Multi-Objective Parameter Tuning), a method that uses a well-known multi-objective optimization algorithm (NSGA-II) to find a front of non-dominated parameter sets which produce good results according to these two metrics. By doing so, we can provide three kinds of results: (i) a method that is able to adapt parameters to a single function, (ii) a comparison between Differential Evolution (DE) and Particle Swarm Optimization (PSO) that takes into consideration both precision and speed, and (iii) an insight into how parameters of DE and PSO affect the performance of these EAs on different benchmark functions
    corecore