233 research outputs found

    Lexicase Selection Outperforms Previous Strategies for Incremental Evolution of Virtual Creature Controllers

    Get PDF
    Evolving robust behaviors for robots has proven to be a challenging problem. Determining how to optimize behavior for a specific instance, while also realizing behaviors that generalize to variations on the problem often requires highly customized algorithms and problem-specific tuning of the evolutionary platform. Algorithms that can realize robust, generalized behavior without this customization are therefore highly desirable. In this paper, we examine the Lexicase selection algorithm as a possible general algorithm for a wall crossing robot task. Previous work has resulted in specialized strategies to evolve robust behaviors for this task. Here, we show that Lexicase selection is not only competitive with these strategies but after parameter tuning, actually exceeds the performance of the specialized algorithms

    Evolving a Behavioral Repertoire for a Walking Robot

    Full text link
    Numerous algorithms have been proposed to allow legged robots to learn to walk. However, the vast majority of these algorithms is devised to learn to walk in a straight line, which is not sufficient to accomplish any real-world mission. Here we introduce the Transferability-based Behavioral Repertoire Evolution algorithm (TBR-Evolution), a novel evolutionary algorithm that simultaneously discovers several hundreds of simple walking controllers, one for each possible direction. By taking advantage of solutions that are usually discarded by evolutionary processes, TBR-Evolution is substantially faster than independently evolving each controller. Our technique relies on two methods: (1) novelty search with local competition, which searches for both high-performing and diverse solutions, and (2) the transferability approach, which com-bines simulations and real tests to evolve controllers for a physical robot. We evaluate this new technique on a hexapod robot. Results show that with only a few dozen short experiments performed on the robot, the algorithm learns a repertoire of con-trollers that allows the robot to reach every point in its reachable space. Overall, TBR-Evolution opens a new kind of learning algorithm that simultaneously optimizes all the achievable behaviors of a robot.Comment: 33 pages; Evolutionary Computation Journal 201

    Exploring New Horizons in Evolutionary Design of Robots

    Get PDF
    International audienceThis introduction paper to the 2009 IROS workshop “Exploring new horizons in Evolutionary Design of Robots” considers the field of Evolutionary Robotics (ER) from the perspective of its potential users: roboticists. The core hypothesis motivating this field of research will be discussed, as well as the potential use of ER in a robot design process. Three main aspects of ER will be presented: (a) ER as an automatic parameter tuning procedure, which is the most mature application and is used to solve real robotics problem, (b) evolutionary-aided design, which may benefit the designer as an efficient tool to build robotic systems and (c) automatic synthesis, which corresponds to the automatic design of a mechatronic device. Critical issues will also be presented as well as current trends and pespectives in ER

    Higher coordination with less control - A result of information maximization in the sensorimotor loop

    Full text link
    This work presents a novel learning method in the context of embodied artificial intelligence and self-organization, which has as few assumptions and restrictions as possible about the world and the underlying model. The learning rule is derived from the principle of maximizing the predictive information in the sensorimotor loop. It is evaluated on robot chains of varying length with individually controlled, non-communicating segments. The comparison of the results shows that maximizing the predictive information per wheel leads to a higher coordinated behavior of the physically connected robots compared to a maximization per robot. Another focus of this paper is the analysis of the effect of the robot chain length on the overall behavior of the robots. It will be shown that longer chains with less capable controllers outperform those of shorter length and more complex controllers. The reason is found and discussed in the information-geometric interpretation of the learning process

    When Specialists Transition to Generalists: Evolutionary Pressure in Lexicase Selection

    Get PDF
    Generalized behavior is a long standing goal for evolutionary robotics. Behaviors for a given task should be robust to perturbation and capable of operating across a variety of environments. We have previously shown that Lexicase selection evolves high-performing individuals in a semi-generalized wall crossing task–i.e., where the task is broadly the same, but there is variation between individual instances. Further work has identified effective parameter values for Lexicase selection in this domain but other factors affecting and explaining performance remain to be identified. In this paper, we expand our prior investigations, examining populations over evolutionary time exploring other factors that might lead to generalized behavior. Results show that genomic clusters do not correspond to performance, indicating that clusters of specialists do not form within the population. While early individuals gain a foothold in the selection process by specializing on a few wall heights, successful populations are ultimately pressured towards generalized behavior. Finally, we find that this transition from specialists to generalists also leads to an increase in tiebreaks, a mechanism in Lexicase, during selection providing a metric to assess the performance of individual replicates

    Behavioral repertoire learning in robotics

    Get PDF
    Behavioral Repertoire Learning in Robotics Antoine Cully ISIR, UniversitĂ© Pierre et Marie Curie-Paris 6, CNRS UMR 7222 4 place Jussieu, F-75252, Paris Cedex 05, France [email protected] Jean-Baptiste Mouret ISIR, UniversitĂ© Pierre et Marie Curie-Paris 6, CNRS UMR 7222 4 place Jussieu, F-75252, Paris Cedex 05, France [email protected] ABSTRACT Learning in robotics typically involves choosing a simple goal (e.g. walking) and assessing the performance of each con- troller with regard to this task (e.g. walking speed). How- ever, learning advanced, input-driven controllers (e.g. walk- ing in each direction) requires testing each controller on a large sample of the possible input signals. This costly pro- cess makes difficult to learn useful low-level controllers in robotics. Here we introduce BR-Evolution, a new evolutionary learn- ing technique that generates a behavioral repertoire by tak- ing advantage of the candidate solutions that are usually discarded. Instead of evolving a single, general controller, BR-evolution thus evolves a collection of simple controllers, one for each variant of the target behavior; to distinguish similar controllers, it uses a performance objective that al- lows it to produce a collection of diverse but high-performing behaviors. We evaluated this new technique by evolving gait controllers for a simulated hexapod robot. Results show that a single run of the EA quickly finds a collection of controllers that allows the robot to reach each point of the reachable space. Overall, BR-Evolution opens a new kind of learning algorithm that simultaneously optimizes all the achievable behaviors of a robot

    Objective Sampling Strategies for Generalized Locomotion Behavior with Lexicase Selection

    Get PDF
    Controllers capable of exhibiting multiple behaviors is a longstanding goal in artificial life. Evolutionary robotics approaches have demonstrated effective optimization of robotic controllers, realizing single behaviors in a variety of domains. However, evolving multiple behaviors in one controller remains an outstanding challenge. Many objective selection algorithms are a potential solution as they are capable of optimizing across tens or hundreds of objectives. In this study, we use Lexicase selection evolving animats capable of both wall crossing and turn/seek behaviors. Our investigation focuses on the objective sampling strategy during selection to balance performance across the two primary tasks. Results show that the sampling strategy does not significantly alter performance, but the number of evaluations required varies significantly across strategies
    • …
    corecore