79 research outputs found

    Self-adaptation of mutation distribution in evolution strategies for dynamic optimization problems

    Get PDF
    Copyright @ IOS Press. All Rights Reserved.Evolution strategies with q-Gaussian mutation, which allows the self-adaptation of the mutation distribution shape, is proposed for dynamic optimization problems in this paper. In the proposed method, a real parameter q, which allows to smoothly control the shape of the mutation distribution, is encoded in the chromosome of the individuals and is allowed to evolve. In the experimental study, the q-Gaussian mutation is compared to Gaussian and Cauchy mutation on experiments generated from the simulation of evolutionary robots and on dynamic optimization problems generated by the Moving Peaks generator

    Efficient Hill Climber for Constrained Pseudo-Boolean Optimization Problems

    Get PDF
    Efficient hill climbers have been recently proposed for single- and multi-objective pseudo-Boolean optimization problems. For kk-bounded pseudo-Boolean functions where each variable appears in at most a constant number of subfunctions, it has been theoretically proven that the neighborhood of a solution can be explored in constant time. These hill climbers, combined with a high-level exploration strategy, have shown to improve state of the art methods in experimental studies and open the door to the so-called Gray Box Optimization, where part, but not all, of the details of the objective functions are used to better explore the search space. One important limitation of all the previous proposals is that they can only be applied to unconstrained pseudo-Boolean optimization problems. In this work, we address the constrained case for multi-objective kk-bounded pseudo-Boolean optimization problems. We find that adding constraints to the pseudo-Boolean problem has a linear computational cost in the hill climber.Universidad de Málaga. Campus de Excelencia Internacional Andalucía Tech

    Efficient Hill Climber for Multi-Objective Pseudo-Boolean Optimization

    Get PDF
    Chicano, F., Whitley D., & Tinós R. (2016). Efficient Hill Climber for Multi-Objective Pseudo-Boolean Optimization. 16th European Conference on Evolutionary Computation for Combinatorial Optimization (LNCS 9595), pp. 88-103Local search algorithms and iterated local search algorithms are a basic technique. Local search can be a stand-alone search method, but it can also be hybridized with evolutionary algorithms. Recently, it has been shown that it is possible to identify improving moves in Hamming neighborhoods for k-bounded pseudo-Boolean optimization problems in constant time. This means that local search does not need to enumerate neighborhoods to find improving moves. It also means that evolutionary algorithms do not need to use random mutation as a operator, except perhaps as a way to escape local optima. In this paper, we show how improving moves can be identified in constant time for multiobjective problems that are expressed as k-bounded pseudo-Boolean functions. In particular, multiobjective forms of NK Landscapes and Mk Landscapes are considered.Universidad de Málaga. Campus de Excelencia Internacional Andalucía Tech. Fulbright program, Ministerio de Educación (CAS12/00274), Ministerio de Economía y Competitividad (TIN2014-57341-R), Air Force Office of Scientific Research, Air Force Materiel Command, USAF (FA9550-11-1-0088), FAPESP (2015/06462-1) and CNPq

    Feature learning in feature-sample networks using multi-objective optimization

    Full text link
    Data and knowledge representation are fundamental concepts in machine learning. The quality of the representation impacts the performance of the learning model directly. Feature learning transforms or enhances raw data to structures that are effectively exploited by those models. In recent years, several works have been using complex networks for data representation and analysis. However, no feature learning method has been proposed for such category of techniques. Here, we present an unsupervised feature learning mechanism that works on datasets with binary features. First, the dataset is mapped into a feature--sample network. Then, a multi-objective optimization process selects a set of new vertices to produce an enhanced version of the network. The new features depend on a nonlinear function of a combination of preexisting features. Effectively, the process projects the input data into a higher-dimensional space. To solve the optimization problem, we design two metaheuristics based on the lexicographic genetic algorithm and the improved strength Pareto evolutionary algorithm (SPEA2). We show that the enhanced network contains more information and can be exploited to improve the performance of machine learning methods. The advantages and disadvantages of each optimization strategy are discussed.Comment: 7 pages, 4 figure

    Optimizing One Million Variable NK Landscapes by Hybridizing Deterministic Recombination and Local Search

    Get PDF
    In gray-box optimization, the search algorithms have access to the variable interaction graph (VIG) of the optimization problem. For Mk Landscapes (and NK Landscapes) we can use the VIG to identify an improving solution in the Hamming neighborhood in constant time. In addition, using the VIG, deterministic Partition Crossover is able to explore an exponential number of solutions in a time that is linear in the size of the problem. Both methods have been used in isolation in previous search algorithms. We present two new gray-box algorithms that combine Partition Crossover with highly efficient local search. The best algorithms are able to locate the global optimum on Adjacent NK Landscape instances with one million variables. The algorithms are compared with a state-of-the-art algorithm for pseudo-Boolean optimization: Gray-Box Parameterless Population Pyramid. The results show that the best algorithm is always one combining Partition Crossover and highly efficient local search. But the results also illustrate that the best optimizer differs on Adjacent and Random NK Landscapes.Universidad de Málaga. Campus de Excelencia Internacional Andalucía Tech

    Enhancing partition crossover with articulation points analysis

    Get PDF
    Partition Crossover is a recombination operator for pseudo-Boolean optimization with the ability to explore an exponential number of solutions in linear or square time. It decomposes the objective function as a sum of subfunctions, each one depending on a different set of variables. The decomposition makes it possible to select the best parent for each subfunction independently, and the operator provides the best out of 2q2^q solutions, where qq is the number of subfunctions in the decomposition. These subfunctions are defined over the connected components of the recombination graph: a subgraph of the objective function variable interaction graph containing only the differing variables in the two parents. In this paper, we advance further and propose a new way to increase the number of linearly independent subfunctions by analyzing the articulation points of the recombination graph. These points correspond to variables that, once flipped, increase the number of connected components. The presence of a connected component with an articulation point increases the number of explored solutions by a factor of, at least, 4. We evaluate the new operator using Iterated Local Search combined with Partition Crossover to solve NK Landscapes and MAX-SAT.Universidad de Málaga. Campus de Excelencia Internacional Andalucía Tech. Funding was provided by the Fulbright program, the Spanish Ministry of Education, Culture and Sport (CAS12/00274), the Spanish Ministry of Economy and Competitiveness and FEDER (TIN2014-57341-R and TIN2017-88213-R), the Air Force Office of Scientific Research, (FA9550-11-1-0088), the Leverhulme Trust (RPG-2015-395), the FAPESP (2015/06462-1) and CNPq (304400/2014-9)

    A fault detection and isolation system for cooperative manipulators

    Get PDF
    The problem of fault detection and isolation (FDI) in cooperative manipulators is addressed in this paper. Four FDI procedures are developed to deal with free-swinging joint faults, locked joint faults, incorrectly measured joint position, and incorrectly measured joint velocity. Free-swinging and locked joint faults are isolated via neural networks. For each arm, a Multilayer Perceptron (MLP) is used to reproduce the dynamics of the fault-free robot. The outputs of each MLP are compared to the actual joint velocities in order to generate a residual vector which is then classified by an RBF network. The remaining faults are isolated based on the kinematic constraints imposed on the cooperative system. Results obtained via simulations and via an actual cooperative manipulator robot are presented

    A New Generalized Partition Crossover for the Traveling Salesman Problem: Tunneling Between Local Optima

    Get PDF
    Generalized Partition Crossover (GPX) is a deterministic recombination operator developed for the Traveling Salesman Problem. Partition crossover operators return the best of 2 k reachable offspring, where k is the number of recombining components. This paper introduces a new GPX2 operator, which finds more recombining components than GPX or Iterative Partial Transcription (IPT). We also show that GPX2 has O(n) runtime complexity, while also introducing new enhancements to reduce the execution time of GPX2. Finally, we experimentally demonstrate the efficiency of GPX2 when it is used to improve solutions found by multi-trial Lin-Kernighan-Helsgaum (LKH) algorithm. Significant improvements in performance are documented on large (n > 5000) and very large (n = 100, 000) instances of the Traveling Salesman Problem

    Use of the q-Gaussian mutation in evolutionary algorithms

    Get PDF
    Copyright @ Springer-Verlag 2010.This paper proposes the use of the q-Gaussian mutation with self-adaptation of the shape of the mutation distribution in evolutionary algorithms. The shape of the q-Gaussian mutation distribution is controlled by a real parameter q. In the proposed method, the real parameter q of the q-Gaussian mutation is encoded in the chromosome of individuals and hence is allowed to evolve during the evolutionary process. In order to test the new mutation operator, evolution strategy and evolutionary programming algorithms with self-adapted q-Gaussian mutation generated from anisotropic and isotropic distributions are presented. The theoretical analysis of the q-Gaussian mutation is also provided. In the experimental study, the q-Gaussian mutation is compared to Gaussian and Cauchy mutations in the optimization of a set of test functions. Experimental results show the efficiency of the proposed method of self-adapting the mutation distribution in evolutionary algorithms.This work was supported in part by FAPESP and CNPq in Brazil and in part by the Engineering and Physical Sciences Research Council (EPSRC) of the UK under Grant EP/E060722/1 and Grant EP/E060722/2
    corecore