972 research outputs found
Shedding light on social learning
Culture involves the origination and transmission of ideas, but the
conditions in which culture can emerge and evolve are unclear. We constructed
and studied a highly simplified neural-network model of these processes. In
this model ideas originate by individual learning from the environment and are
transmitted by communication between individuals. Individuals (or "agents")
comprise a single neuron which receives structured data from the environment
via plastic synaptic connections. The data are generated in the simplest
possible way: linear mixing of independently fluctuating sources and the goal
of learning is to unmix the data. To make this problem tractable we assume that
at least one of the sources fluctuates in a nonGaussian manner. Linear mixing
creates structure in the data, and agents attempt to learn (from the data and
possibly from other individuals) synaptic weights that will unmix, i.e., to
"understand" the agent's world. For a variety of reasons even this goal can be
difficult for a single agent to achieve; we studied one particular type of
difficulty (created by imperfection in synaptic plasticity), though our
conclusions should carry over to many other types of difficulty. We previously
studied whether a small population of communicating agents, learning from each
other, could more easily learn unmixing coefficients than isolated individuals,
learning only from their environment. We found, unsurprisingly, that if agents
learn indiscriminately from any other agent (whether or not they have learned
good solutions), communication does not enhance understanding. Here we extend
the model slightly, by allowing successful learners to be more effective
teachers, and find that now a population of agents can learn more effectively
than isolated individuals. We suggest that a key factor in the onset of culture
might be the development of selective learning.Comment: 11 pages 8 figure
A Comprehensive Workflow for General-Purpose Neural Modeling with Highly Configurable Neuromorphic Hardware Systems
In this paper we present a methodological framework that meets novel
requirements emerging from upcoming types of accelerated and highly
configurable neuromorphic hardware systems. We describe in detail a device with
45 million programmable and dynamic synapses that is currently under
development, and we sketch the conceptual challenges that arise from taking
this platform into operation. More specifically, we aim at the establishment of
this neuromorphic system as a flexible and neuroscientifically valuable
modeling tool that can be used by non-hardware-experts. We consider various
functional aspects to be crucial for this purpose, and we introduce a
consistent workflow with detailed descriptions of all involved modules that
implement the suggested steps: The integration of the hardware interface into
the simulator-independent model description language PyNN; a fully automated
translation between the PyNN domain and appropriate hardware configurations; an
executable specification of the future neuromorphic system that can be
seamlessly integrated into this biology-to-hardware mapping process as a test
bench for all software layers and possible hardware design modifications; an
evaluation scheme that deploys models from a dedicated benchmark library,
compares the results generated by virtual or prototype hardware devices with
reference software simulations and analyzes the differences. The integration of
these components into one hardware-software workflow provides an ecosystem for
ongoing preparative studies that support the hardware design process and
represents the basis for the maturity of the model-to-hardware mapping
software. The functionality and flexibility of the latter is proven with a
variety of experimental results
NASA JSC neural network survey results
A survey of Artificial Neural Systems in support of NASA's (Johnson Space Center) Automatic Perception for Mission Planning and Flight Control Research Program was conducted. Several of the world's leading researchers contributed papers containing their most recent results on artificial neural systems. These papers were broken into categories and descriptive accounts of the results make up a large part of this report. Also included is material on sources of information on artificial neural systems such as books, technical reports, software tools, etc
Organization, Maturation, and Plasticity of Multisensory Integration: Insights from Computational Modeling Studies
In this paper, we present two neural network models â devoted to two specific and widely investigated aspects of multisensory integration â in order to evidence the potentialities of computational models to gain insight into the neural mechanisms underlying organization, development, and plasticity of multisensory integration in the brain. The first model considers visualâauditory interaction in a midbrain structure named superior colliculus (SC). The model is able to reproduce and explain the main physiological features of multisensory integration in SC neurons and to describe how SC integrative capability â not present at birth â develops gradually during postnatal life depending on sensory experience with cross-modal stimuli. The second model tackles the problem of how tactile stimuli on a body part and visual (or auditory) stimuli close to the same body part are integrated in multimodal parietal neurons to form the perception of peripersonal (i.e., near) space. The model investigates how the extension of peripersonal space â where multimodal integration occurs â may be modified by experience such as use of a tool to interact with the far space. The utility of the modeling approach relies on several aspects: (i) The two models, although devoted to different problems and simulating different brain regions, share some common mechanisms (lateral inhibition and excitation, non-linear neuron characteristics, recurrent connections, competition, Hebbian rules of potentiation and depression) that may govern more generally the fusion of senses in the brain, and the learning and plasticity of multisensory integration. (ii) The models may help interpretation of behavioral and psychophysical responses in terms of neural activity and synaptic connections. (iii) The models can make testable predictions that can help guiding future experiments in order to validate, reject, or modify the main assumptions
- âŠ