4,496 research outputs found

    Bayesian Reinforcement Learning via Deep, Sparse Sampling

    Full text link
    We address the problem of Bayesian reinforcement learning using efficient model-based online planning. We propose an optimism-free Bayes-adaptive algorithm to induce deeper and sparser exploration with a theoretical bound on its performance relative to the Bayes optimal policy, with a lower computational complexity. The main novelty is the use of a candidate policy generator, to generate long-term options in the planning tree (over beliefs), which allows us to create much sparser and deeper trees. Experimental results on different environments show that in comparison to the state-of-the-art, our algorithm is both computationally more efficient, and obtains significantly higher reward in discrete environments.Comment: Published in AISTATS 202

    Sparsely Aggregated Convolutional Networks

    Full text link
    We explore a key architectural aspect of deep convolutional neural networks: the pattern of internal skip connections used to aggregate outputs of earlier layers for consumption by deeper layers. Such aggregation is critical to facilitate training of very deep networks in an end-to-end manner. This is a primary reason for the widespread adoption of residual networks, which aggregate outputs via cumulative summation. While subsequent works investigate alternative aggregation operations (e.g. concatenation), we focus on an orthogonal question: which outputs to aggregate at a particular point in the network. We propose a new internal connection structure which aggregates only a sparse set of previous outputs at any given depth. Our experiments demonstrate this simple design change offers superior performance with fewer parameters and lower computational requirements. Moreover, we show that sparse aggregation allows networks to scale more robustly to 1000+ layers, thereby opening future avenues for training long-running visual processes.Comment: Accepted to ECCV 201

    Systematic revision of the american taxa belonging to the genera Alloblackburneus Bordat, 2009, and Blackburneus Schmidt, 1913, with description of seven new species (Coleoptera: Scarabaeidae: Aphodiinae)

    Get PDF
    The American species belonging to the genera Alloblackburneus Bordat and Blackburneus Schmidt (Coleoptera: Scarabaeidae: Aphodiinae) are redescribed and figured. Seven new species are described: Alloblackburneus guadalajarae, A. ibanezbernali; Blackburneus amazonicus, B. sanfilippoi, B. surinamensis, B. teposcolulaensis, B. thomasi. The neotype of Scarabaeus rubeolus Palisot de Beauvois, 1809 is designated. The lectotype of Blackburneus argentinensis (Schmidt, 1909) and of Blackburneus laxepunctatus (Schmidt, 1910) are designated. The following new combinations are proposed: Alloblackburneus aegrotus (Horn, 1870); Alloblackburneus cavidomus (Brown, 1927); Alloblackburneus cynomysi (Brown, 1927); Alloblackburneus fordi (Gordon, 1974); Alloblackburneus geomysi (Cartwright, 1939); Alloblackburneus lentus (Horn, 1870); Alloblackburneus rubeolus (Palisot de Beauvois, 1805); Alloblackburneus saylori (Hinton, 1934); Alloblackburneus tenuistriatus (Horn, 1887); Alloblackburneus troglodytes (Hubbard, 1894)

    Sample Efficient Bayesian Reinforcement Learning

    Get PDF
    Artificial Intelligence (AI) has been an active field of research for over a century now. The research field of AI may be grouped into various tasks that are expected from an intelligent agent; two major ones being learning & inference and planning. The act of storing new knowledge is known as learning while inference refers to the act to extracting conclusions given agent’s limited knowledge base. They are tightly knit by the design of its knowledge base. The process of deciding long-term actions or plans given its current knowledge is called planning.Reinforcement Learning (RL) brings together these two tasks by posing a seemingly benign question “How to act optimally in an unknown environment?”. This requires the agent to learn about its environment as well as plan actions given its current knowledge about it. In RL, the environment can be represented by a mathematical model and we associate an intrinsic value to the actions that the agent may choose.In this thesis, we present a novel Bayesian algorithm for the problem of RL. Bayesian RL is a widely explored area of research but is constrained by scalability and performance issues. We provide first steps towards rigorous analysis of these types of algorithms. Bayesian algorithms are characterized by the belief that they maintain over their unknowns; which is updated based on the collected evidence. This is different from the traditional approach in RL in terms of problem formulation and formal guarantees. Our novel algorithm combines aspects of planning and learning due to its inherent Bayesian formulation. It does so in a more scalable fashion, with formal PAC guarantees. We also give insights on the application of Bayesian framework for the estimation of model and value, in a joint work on Bayesian backward induction for RL
    • …
    corecore