775 research outputs found

    AI of Brain and Cognitive Sciences: From the Perspective of First Principles

    Full text link
    Nowadays, we have witnessed the great success of AI in various applications, including image classification, game playing, protein structure analysis, language translation, and content generation. Despite these powerful applications, there are still many tasks in our daily life that are rather simple to humans but pose great challenges to AI. These include image and language understanding, few-shot learning, abstract concepts, and low-energy cost computing. Thus, learning from the brain is still a promising way that can shed light on the development of next-generation AI. The brain is arguably the only known intelligent machine in the universe, which is the product of evolution for animals surviving in the natural environment. At the behavior level, psychology and cognitive sciences have demonstrated that human and animal brains can execute very intelligent high-level cognitive functions. At the structure level, cognitive and computational neurosciences have unveiled that the brain has extremely complicated but elegant network forms to support its functions. Over years, people are gathering knowledge about the structure and functions of the brain, and this process is accelerating recently along with the initiation of giant brain projects worldwide. Here, we argue that the general principles of brain functions are the most valuable things to inspire the development of AI. These general principles are the standard rules of the brain extracting, representing, manipulating, and retrieving information, and here we call them the first principles of the brain. This paper collects six such first principles. They are attractor network, criticality, random network, sparse coding, relational memory, and perceptual learning. On each topic, we review its biological background, fundamental property, potential application to AI, and future development.Comment: 59 pages, 5 figures, review articl

    Layered architecture for quantum computing

    Full text link
    We develop a layered quantum computer architecture, which is a systematic framework for tackling the individual challenges of developing a quantum computer while constructing a cohesive device design. We discuss many of the prominent techniques for implementing circuit-model quantum computing and introduce several new methods, with an emphasis on employing surface code quantum error correction. In doing so, we propose a new quantum computer architecture based on optical control of quantum dots. The timescales of physical hardware operations and logical, error-corrected quantum gates differ by several orders of magnitude. By dividing functionality into layers, we can design and analyze subsystems independently, demonstrating the value of our layered architectural approach. Using this concrete hardware platform, we provide resource analysis for executing fault-tolerant quantum algorithms for integer factoring and quantum simulation, finding that the quantum dot architecture we study could solve such problems on the timescale of days.Comment: 27 pages, 20 figure

    Prefrontal rhythms for cognitive control

    Get PDF
    Goal-directed behavior requires flexible selection among action plans and updating behavioral strategies when they fail to achieve desired goals. Lateral prefrontal cortex (LPFC) is implicated in the execution of behavior-guiding rule-based cognitive control while anterior cingulate cortex (ACC) is implicated in monitoring processes and updating rules. Rule-based cognitive control requires selective processing while process monitoring benefits from combinatorial processing. I used a combination of computational and experimental methods to investigate how network oscillations and neuronal heterogeneity contribute to cognitive control through their effects on selective versus combinatorial processing modes in LPFC and ACC. First, I adapted an existing LPFC model to explore input frequency- and coherence-based output selection mechanisms for flexible routing of rate-coded signals. I show that the oscillatory states of input encoding populations can exhibit a stronger influence over downstream competition than their activity levels. This enables an output driven by a weaker resonant input signal to suppress lower-frequency competing responses to stronger, less resonant (though possibly higher-frequency) input signals. While signals are encoded in population firing rates, output selection and signal routing can be governed independently by the frequency and coherence of oscillatory inputs and their correspondence with output resonant properties. Flexible response selection and gating can be achieved by oscillatory state control mechanisms operating on input encoding populations. These dynamic mechanisms enable experimentally-observed LPFC beta and gamma oscillations to flexibly govern the selection and gating of rate-coded signals for downstream read-out. Furthermore, I demonstrate how differential drives to distinct interneuron populations can switch working memory representations between asynchronous and oscillatory states that support rule-based selection. Next, I analyzed physiological data from the LeBeau laboratory and built a de novo model constrained by the biological data. Experimental data demonstrated that fast network oscillations at both the beta- and gamma frequency bands could be elicited in vitro in ACC and neurons exhibited a wide range of intrinsic properties. Computational modeling of the ACC network revealed that the frequency of network oscillation generated was dependent upon the time course of inhibition. Principal cell heterogeneity broadened the range of frequencies generated by the model network. In addition, with different frequency inputs to two neuronal assemblies, heterogeneity decreased competition and increased spike coherence between the networks thus conferring a combinatorial advantage to the network. These findings suggest that oscillating neuronal populations can support either response selection (routing), or combination, depending on the interplay between the kinetics of synaptic inhibition and the degree of heterogeneity of principal cell intrinsic conductances. Such differences may support functional differences between the roles of LPFC and ACC in cognitive control

    Enabling and Understanding Failure of Engineering Structures Using the Technique of Cohesive Elements

    Get PDF
    In this paper, we describe a cohesive zone model for the prediction of failure of engineering solids and/or structures. A damage evolution law is incorporated into a three-dimensional, exponential cohesive law to account for material degradation under the influence of cyclic loading. This cohesive zone model is implemented in the finite element software ABAQUS through a user defined subroutine. The irreversibility of the cohesive zone model is first verified and subsequently applied for studying cyclic crack growth in specimens experiencing different modes of fracture and/or failure. The crack growth behavior to include both crack initiation and crack propagation becomes a natural outcome of the numerical simulation. Numerical examples suggest that the irreversible cohesive zone model can serve as an efficient tool to predict fatigue crack growth. Key issues such as crack path deviation, convergence and mesh dependency are also discussed

    Advances in Reinforcement Learning

    Get PDF
    Reinforcement Learning (RL) is a very dynamic area in terms of theory and application. This book brings together many different aspects of the current research on several fields associated to RL which has been growing rapidly, producing a wide variety of learning algorithms for different applications. Based on 24 Chapters, it covers a very broad variety of topics in RL and their application in autonomous systems. A set of chapters in this book provide a general overview of RL while other chapters focus mostly on the applications of RL paradigms: Game Theory, Multi-Agent Theory, Robotic, Networking Technologies, Vehicular Navigation, Medicine and Industrial Logistic

    Many-core architectures with time predictable execution Support for hard real-time applications

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2013.Cataloged from PDF version of thesis.Includes bibliographical references (p. 183-193).Hybrid control systems are a growing domain of application. They are pervasive and their complexity is increasing rapidly. Distributed control systems for future "Intelligent Grid" and renewable energy generation systems are demanding high-performance, hard real-time computation, and more programmability. General-purpose computer systems are primarily designed to process data and not to interact with physical processes as required by these systems. Generic general-purpose architectures even with the use of real-time operating systems fail to meet the hard realtime constraints of hybrid system dynamics. ASIC, FPGA, or traditional embedded design approaches to these systems often result in expensive, complicated systems that are hard to program, reuse, or maintain. In this thesis, we propose a domain-specific architecture template targeting hybrid control system applications. Using power electronics control applications, we present new modeling techniques, synthesis methodologies, and a parameterizable computer architecture for these large distributed control systems. We propose a new system modeling approach, called Adaptive Hybrid Automaton, based on previous work in control system theory, that uses a mixed-model abstractions and lends itself well to digital processing. We develop a domain-specific architecture based on this modeling that uses heterogeneous processing units and predictable execution, called MARTHA. We develop a hard real-time aware router architecture to enable deterministic on-chip interconnect network communication. We present several algorithms for scheduling task-based applications onto these types of heterogeneous architectures. We create Heracles, an open-source, functional, parameterized, synthesizable many-core system design toolkit, that can be used to explore future multi/many-core processors with different topologies, routing schemes, processing elements or cores, and memory system organizations. Using the Heracles design tool we build a prototype of the proposed architecture using a state-of-the-art FPGA-based platform, and deploy and test it in actual physical power electronics systems. We develop and release an open-source, small representative set of power electronics system applications that can be used for hard real-time application benchmarking.by Michel A. Kinsy.Ph.D
    • …
    corecore