60 research outputs found

    Fast deep reinforcement learning using online adjustments from the past

    Get PDF
    We propose Ephemeral Value Adjusments (EVA): a means of allowing deep reinforcement learning agents to rapidly adapt to experience in their replay buffer. EVA shifts the value predicted by a neural network with an estimate of the value function found by planning over experience tuples from the replay buffer near the current state. EVA combines a number of recent ideas around combining episodic memory-like structures into reinforcement learning agents: slot-based storage, content-based retrieval, and memory-based planning. We show that EVAis performant on a demonstration task and Atari games.Comment: Accepted at NIPS 201

    A graphical programming interface for a children's constructionist learning environment

    Get PDF
    Thesis (M.Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1998.Includes bibliographical references (p. 41).by Andrew C. Cheng.M.Eng

    Model-based reinforcement learning: A survey

    Get PDF
    Reinforcement learning is an important branch of machine learning and artificial intelligence. Compared with traditional reinforcement learning, model-based reinforcement learning obtains the action of the next state by the model that has been learned, and then optimizes the policy, which greatly improves data efficiency. Based on the present status of research on model-based reinforcement learning at home and abroad, this paper comprehensively reviews the key techniques of model-based reinforcement learning, summarizes the characteristics, advantages and defects of each technology, and analyzes the application of model-based reinforcement learning in games, robotics and brain science

    The Daily Egyptian, March 09, 1988

    Get PDF

    The Daily Egyptian, March 09, 1988

    Get PDF

    Natural Curiosity

    Get PDF
    Curiosity is evident in humans of all sorts from early infancy, and it has also been said to appear in a wide range of other animals, including monkeys, birds, rats, and octopuses. The classical definition of curiosity as an intrinsic desire for knowledge may seem inapplicable to animal curiosity: one might wonder how and indeed whether a rat could have such a fancy desire. Even if rats must learn many things to survive, one might expect their learning must be driven by simpler incentives, such as hunger. One might also wonder what proximal signals could guide animals towards knowledge itself, or how something as abstract as knowledge could ever be a motivational target for an unreflective animal. Taking a cue from recent work in reinforcement learning, I argue that surprise functions as a reward signal for the curious animal, and then show how this amounts to a desire for knowledge gain, where knowledge is conceived of as a cognitive adaptation to reality. This adaptation results in a mental state whose existence depends essentially on the truth of its contents, a factive mental state. Curious creatures benefit from an interaction between the prediction-error correction processes of basic learning and the active surprise-seeking force of their curiosity. This internally adversarial interaction accelerates knowledge gain in ways that are very helpful for agents with the restrictions of biological creatures, in environments with the complexity of our natural world

    The Chronicle [October 9, 1979]

    Get PDF
    The Chronicle, October 9, 1979https://repository.stcloudstate.edu/chron/3142/thumbnail.jp
    corecore