950 research outputs found

    Computational Theories of Curiosity-Driven Learning

    Get PDF
    What are the functions of curiosity? What are the mechanisms of curiosity-driven learning? We approach these questions about the living using concepts and tools from machine learning and developmental robotics. We argue that curiosity-driven learning enables organisms to make discoveries to solve complex problems with rare or deceptive rewards. By fostering exploration and discovery of a diversity of behavioural skills, and ignoring these rewards, curiosity can be efficient to bootstrap learning when there is no information, or deceptive information, about local improvement towards these problems. We also explain the key role of curiosity for efficient learning of world models. We review both normative and heuristic computational frameworks used to understand the mechanisms of curiosity in humans, conceptualizing the child as a sense-making organism. These frameworks enable us to discuss the bi-directional causal links between curiosity and learning, and to provide new hypotheses about the fundamental role of curiosity in self-organizing developmental structures through curriculum learning. We present various developmental robotics experiments that study these mechanisms in action, both supporting these hypotheses to understand better curiosity in humans and opening new research avenues in machine learning and artificial intelligence. Finally, we discuss challenges for the design of experimental paradigms for studying curiosity in psychology and cognitive neuroscience. Keywords: Curiosity, intrinsic motivation, lifelong learning, predictions, world model, rewards, free-energy principle, learning progress, machine learning, AI, developmental robotics, development, curriculum learning, self-organization.Comment: To appear in "The New Science of Curiosity", ed. G. Gordon, Nova Science Publisher

    TOWARDS THE GROUNDING OF ABSTRACT CATEGORIES IN COGNITIVE ROBOTS

    Get PDF
    The grounding of language in humanoid robots is a fundamental problem, especially in social scenarios which involve the interaction of robots with human beings. Indeed, natural language represents the most natural interface for humans to interact and exchange information about concrete entities like KNIFE, HAMMER and abstract concepts such as MAKE, USE. This research domain is very important not only for the advances that it can produce in the design of human-robot communication systems, but also for the implication that it can have on cognitive science. Abstract words are used in daily conversations among people to describe events and situations that occur in the environment. Many scholars have suggested that the distinction between concrete and abstract words is a continuum according to which all entities can be varied in their level of abstractness. The work presented herein aimed to ground abstract concepts, similarly to concrete ones, in perception and action systems. This permitted to investigate how different behavioural and cognitive capabilities can be integrated in a humanoid robot in order to bootstrap the development of higher-order skills such as the acquisition of abstract words. To this end, three neuro-robotics models were implemented. The first neuro-robotics experiment consisted in training a humanoid robot to perform a set of motor primitives (e.g. PUSH, PULL, etc.) that hierarchically combined led to the acquisition of higher-order words (e.g. ACCEPT, REJECT). The implementation of this model, based on a feed-forward artificial neural networks, permitted the assessment of the training methodology adopted for the grounding of language in humanoid robots. In the second experiment, the architecture used for carrying out the first study was reimplemented employing recurrent artificial neural networks that enabled the temporal specification of the action primitives to be executed by the robot. This permitted to increase the combinations of actions that can be taught to the robot for the generation of more complex movements. For the third experiment, a model based on recurrent neural networks that integrated multi-modal inputs (i.e. language, vision and proprioception) was implemented for the grounding of abstract action words (e.g. USE, MAKE). Abstract representations of actions ("one-hot" encoding) used in the other two experiments, were replaced with the joints values recorded from the iCub robot sensors. Experimental results showed that motor primitives have different activation patterns according to the action's sequence in which they are embedded. Furthermore, the performed simulations suggested that the acquisition of concepts related to abstract action words requires the reactivation of similar internal representations activated during the acquisition of the basic concepts, directly grounded in perceptual and sensorimotor knowledge, contained in the hierarchical structure of the words used to ground the abstract action words.This study was financed by the EU project RobotDoC (235065) from the Seventh Framework Programme (FP7), Marie Curie Actions Initial Training Network

    Final report key contents: main results accomplished by the EU-Funded project IM-CLeVeR - Intrinsically Motivated Cumulative Learning Versatile Robots

    Get PDF
    This document has the goal of presenting the main scientific and technological achievements of the project IM-CLeVeR. The document is organised as follows: 1. Project executive summary: a brief overview of the project vision, objectives and keywords. 2. Beneficiaries of the project and contacts: list of Teams (partners) of the project, Team Leaders and contacts. 3. Project context and objectives: the vision of the project and its overall objectives 4. Overview of work performed and main results achieved: a one page overview of the main results of the project 5. Overview of main results per partner: a bullet-point list of main results per partners 6. Main achievements in detail, per partner: a throughout explanation of the main results per partner (but including collaboration work), with also reference to the main publications supporting them

    PCT and beyond: toward a computational framework for ‘intelligent’ communicative systems

    No full text
    Recent years have witnessed increasing interest in ‘intelligent’ autonomous machines such as robots. However, there is a long way to go before autonomous systems reach the level of capabilities required for even the simplest of tasks involving human-robot interaction - especially if it involves communicative behavior such as speech and language. The field of Artificial Intelligence (AI) has made great strides in these areas, and has graduated from high-level rule-based paradigms to embodied architectures whose operations are grounded in real physical environments. What is still missing, however, is an overarching theory of intelligent communicative behavior that informs system-level design decisions. This chapter introduces a framework that extends the principles of Perceptual Control Theory (PCT) toward a remarkably symmetric architecture for a needs-driven communicative agent. It is concluded that, if behavior is the control of perception (the central tenet of PCT), then perception (for communicative agents) is the simulation of behavior

    Intelligent systems: towards a new synthetic agenda

    Get PDF

    World model learning and inference

    Get PDF
    Understanding information processing in the brain-and creating general-purpose artificial intelligence-are long-standing aspirations of scientists and engineers worldwide. The distinctive features of human intelligence are high-level cognition and control in various interactions with the world including the self, which are not defined in advance and are vary over time. The challenge of building human-like intelligent machines, as well as progress in brain science and behavioural analyses, robotics, and their associated theoretical formalisations, speaks to the importance of the world-model learning and inference. In this article, after briefly surveying the history and challenges of internal model learning and probabilistic learning, we introduce the free energy principle, which provides a useful framework within which to consider neuronal computation and probabilistic world models. Next, we showcase examples of human behaviour and cognition explained under that principle. We then describe symbol emergence in the context of probabilistic modelling, as a topic at the frontiers of cognitive robotics. Lastly, we review recent progress in creating human-like intelligence by using novel probabilistic programming languages. The striking consensus that emerges from these studies is that probabilistic descriptions of learning and inference are powerful and effective ways to create human-like artificial intelligent machines and to understand intelligence in the context of how humans interact with their world

    The use of emotions in the implementation of various types of learning in a cognitive agent

    Get PDF
    Les tuteurs professionnels humains sont capables de prendre en considération des événements du passé et du présent et ont une capacité d'adaptation en fonction d'événements sociaux. Afin d'être considéré comme une technologie valable pour l'amélioration de l'apprentissage humain, un agent cognitif artificiel devrait pouvoir faire de même. Puisque les environnements dynamiques sont en constante évolution, un agent cognitif doit pareillement évoluer et s'adapter aux modifications structurales et aux phénomènes nouveaux. Par conséquent, l'agent cognitif idéal devrait posséder des capacités d'apprentissage similaires à celles que l'on retrouve chez l'être humain ; l'apprentissage émotif, l'apprentissage épisodique, l'apprentissage procédural, et l'apprentissage causal. Cette thèse contribue à l'amélioration des architectures d'agents cognitifs. Elle propose 1) une méthode d'intégration des émotions inspirée du fonctionnement du cerveau; et 2) un ensemble de méthodes d'apprentissage (épisodique, causale, etc.) qui tiennent compte de la dimension émotionnelle. Le modèle proposé que nous avons appelé CELTS (Conscious Emotional Learning Tutoring System) est une extension d'un agent cognitif conscient dans le rôle d'un tutoriel intelligent. Il comporte un module de gestion des émotions qui permet d'attribuer des valences émotionnelles positives ou négatives à chaque événement perçu par l'agent. Deux voies de traitement sont prévues : 1) une voie courte qui permet au système de répondre immédiatement à certains événements sans un traitement approfondis, et 2) une voie longue qui intervient lors de tout événement qui exige la volition. Dans cette perspective, la dimension émotionnelle est considérée dans les processus cognitifs de l'agent pour la prise de décision et l'apprentissage. L'apprentissage épisodique dans CELTS est basé sur la théorie du Multiple Trace Memory consolidation qui postule que lorsque l'on perçoit un événement, l'hippocampe fait une première interprétation et un premier apprentissage. Ensuite, l'information acquise est distribuée aux différents cortex. Selon cette théorie, la reconsolidation de la mémoire dépend toujours de l'hippocampe. Pour simuler de tel processus, nous avons utilisé des techniques de fouille de données qui permettent la recherche de motifs séquentiels fréquents dans les données générées durant chaque cycle cognitif. L'apprentissage causal dans CELTS se produit à l'aide de la mémoire épisodique. Il permet de trouver les causes et les effets possibles entre différents événements. Il est mise en œuvre grâce à des algorithmes de recherche de règles d'associations. Les associations établies sont utilisées pour piloter les interventions tutorielles de CELTS et, par le biais des réponses de l'apprenant, pour évaluer les règles causales découvertes. \ud ______________________________________________________________________________ \ud MOTS-CLÉS DE L’AUTEUR : agents cognitifs, émotions, apprentissage épisodique, apprentissage causal

    The active inference approach to ecological perception: general information dynamics for natural and artificial embodied cognition

    Get PDF
    The emerging neurocomputational vision of humans as embodied, ecologically embedded, social agents—who shape and are shaped by their environment—offers a golden opportunity to revisit and revise ideas about the physical and information-theoretic underpinnings of life, mind, and consciousness itself. In particular, the active inference framework (AIF) makes it possible to bridge connections from computational neuroscience and robotics/AI to ecological psychology and phenomenology, revealing common underpinnings and overcoming key limitations. AIF opposes the mechanistic to the reductive, while staying fully grounded in a naturalistic and information-theoretic foundation, using the principle of free energy minimization. The latter provides a theoretical basis for a unified treatment of particles, organisms, and interactive machines, spanning from the inorganic to organic, non-life to life, and natural to artificial agents. We provide a brief introduction to AIF, then explore its implications for evolutionary theory, ecological psychology, embodied phenomenology, and robotics/AI research. We conclude the paper by considering implications for machine consciousness
    • …
    corecore