1,076 research outputs found

    Inference in classifier systems

    Get PDF
    Classifier systems (Css) provide a rich framework for learning and induction, and they have beenı successfully applied in the artificial intelligence literature for some time. In this paper, both theı architecture and the inferential mechanisms in general CSs are reviewed, and a number of limitations and extensions of the basic approach are summarized. A system based on the CS approach that is capable of quantitative data analysis is outlined and some of its peculiarities discussed

    RTCS: a reactive with tags classifier system

    Get PDF
    In this work, a new Classifier System is proposed (CS). The system, a Reactive with Tags Classifier System (RTCS), is able to take into account environmental situations in intermediate decisions. CSs are special production systems, where conditions and actions are codified in order to learn new rules by means of Genetic Algorithms (GA). The RTCS has been designed to generate sequences of actions like the traditional classifier systems, but RTCS also has the capability of chaining rules among different time instants and reacting to new environmental situations, considering the last environmental situation to take a decision. In addition to the capability to react and generate sequences of actions, the design of a new rule codification allows the evolution of groups of specialized rules. This new codification is based on the inclusion of several bits, named tags, in conditions and actions, which evolve by means of GA. RTCS has been tested in robotic navigation. Results show the suitability of this approximation to the navigation problem and the coherence of tag values in rules classification.Publicad

    An enhanced classifier system for autonomous robot navigation in dynamic environments

    Get PDF
    In many cases, a real robot application requires the navigation in dynamic environments. The navigation problem involves two main tasks: to avoid obstacles and to reach a goal. Generally, this problem could be faced considering reactions and sequences of actions. For solving the navigation problem a complete controller, including actions and reactions, is needed. Machine learning techniques has been applied to learn these controllers. Classifier Systems (CS) have proven their ability of continuos learning in these domains. However, CS have some problems in reactive systems. In this paper, a modified CS is proposed to overcome these problems. Two special mechanisms are included in the developed CS to allow the learning of both reactions and sequences of actions. The learning process has been divided in two main tasks: first, the discrimination between a predefined set of rules and second, the discovery of new rules to obtain a successful operation in dynamic environments. Different experiments have been carried out using a mini-robot Khepera to find a generalised solution. The results show the ability of the system to continuous learning and adaptation to new situations.Publicad

    Genetic algorithms for adaptive real-time control in space systems

    Get PDF
    Genetic Algorithms that are used for learning as one way to control the combinational explosion associated with the generation of new rules are discussed. The Genetic Algorithm approach tends to work best when it can be applied to a domain independent knowledge representation. Applications to real time control in space systems are discussed

    A reactive approach to classifier systems

    Get PDF
    IEEE International Conference on Systems, Man, and Cybernetics. San Diego, CA, 11-14 Oct. 1998The navigation problem involves how to reach a goal avoiding obstacles in dynamic environments. This problem can be faced considering reactions and/or sequences of actions. Classifier Systems (CS) have proven their ability of continuous learning, however they have some problems in reactive systems. A modified CS is proposed to overcome these problems. Two special mechanisms are included in the developed CS to allow the learning of both reactions and sequences of actions. This learning process involves two main tasks: first, discriminating between rules and second, the discovery of new rules to obtain a successful operation in dynamic environments. Different experiments have been carried out using a mini-robot Khepera to find a generalized solution. The results show the ability of the system for continuous learning and adaptation to new situations

    Meta-Genetic Programming: Co-evolving the Operators of Variation

    Get PDF
    The standard Genetic Programming approach is augmented by co-evolving the genetic operators. To do this the operators are coded as trees of indefinite length. In order for this technique to work, the language that the operators are defined in must be such that it preserves the variation in the base population. This technique can varied by adding further populations of operators and changing which populations act as operators for others, including itself, thus to provide a framework for a whole set of augmented GP techniques. The technique is tested on the parity problem. The pros and cons of the technique are discussed

    Event-driven Hybrid Classifier Systems and Online Learning for Soccer Game Strategies

    Get PDF
    The field of robot soccer is a useful setting for the study of artificial intelligence and machin

    ZCS redux

    Get PDF
    Learning classifier systems traditionally use genetic algorithms to facilitate rule discovery, where rule fitness is payoff based. Current research has shifted to the use of accuracy-based fitness. This paper re-examines the use of a particular payoff-based learning classifier system - ZCS. By using simple difference equation models of ZCS, we show that this system is capable of optimal performance subject to appropriate parameter settings. This is demonstrated for both single- and multistep tasks. Optimal performance of ZCS in well-known, multistep maze tasks is then presented to support the findings from the models

    Evolutionary Algorithms for Reinforcement Learning

    Full text link
    There are two distinct approaches to solving reinforcement learning problems, namely, searching in value function space and searching in policy space. Temporal difference methods and evolutionary algorithms are well-known examples of these approaches. Kaelbling, Littman and Moore recently provided an informative survey of temporal difference methods. This article focuses on the application of evolutionary algorithms to the reinforcement learning problem, emphasizing alternative policy representations, credit assignment methods, and problem-specific genetic operators. Strengths and weaknesses of the evolutionary approach to reinforcement learning are presented, along with a survey of representative applications

    Learning sequences of rules using classifier systems with tags

    Get PDF
    IEEE International Conference on Systems, Man, and Cybernetics. Tokyo, 12-15 October 1999.The objective of this paper was to obtain an encoding structure that would allow the genetic evolution of rules in such a manner that the number of rules and relationship in a classifier system (CS) would be learnt in the evolution process. For this purpose, an area that allows the definition of rule groups has been entered into the condition and message part of the encoded rules. This area is called internal tag. This term was coined because the system has some similarities with natural processes that take place in certain animal species, where the existence of tags allows them to communicate and recognize each other. Such CS is called a tag classifier system (TCS). The TCS has been tested in the game of draughts and compared with the classical CS. The results show an improving of the CS performance
    • …
    corecore