23,896 research outputs found
Real time unsupervised learning of visual stimuli in neuromorphic VLSI systems
Neuromorphic chips embody computational principles operating in the nervous
system, into microelectronic devices. In this domain it is important to
identify computational primitives that theory and experiments suggest as
generic and reusable cognitive elements. One such element is provided by
attractor dynamics in recurrent networks. Point attractors are equilibrium
states of the dynamics (up to fluctuations), determined by the synaptic
structure of the network; a `basin' of attraction comprises all initial states
leading to a given attractor upon relaxation, hence making attractor dynamics
suitable to implement robust associative memory. The initial network state is
dictated by the stimulus, and relaxation to the attractor state implements the
retrieval of the corresponding memorized prototypical pattern. In a previous
work we demonstrated that a neuromorphic recurrent network of spiking neurons
and suitably chosen, fixed synapses supports attractor dynamics. Here we focus
on learning: activating on-chip synaptic plasticity and using a theory-driven
strategy for choosing network parameters, we show that autonomous learning,
following repeated presentation of simple visual stimuli, shapes a synaptic
connectivity supporting stimulus-selective attractors. Associative memory
develops on chip as the result of the coupled stimulus-driven neural activity
and ensuing synaptic dynamics, with no artificial separation between learning
and retrieval phases.Comment: submitted to Scientific Repor
Dynamical model of sequential spatial memory: winnerless competition of patterns
We introduce a new biologically-motivated model of sequential spatial memory
which is based on the principle of winnerless competition (WLC). We implement
this mechanism in a two-layer neural network structure and present the learning
dynamics which leads to the formation of a WLC network. After learning, the
system is capable of associative retrieval of pre-recorded sequences of spatial
patterns.Comment: 4 pages, submitted to PR
Neuromorphic Computing Applications in Robotics
Deep learning achieves remarkable success through training using massively labeled datasets. However, the high demands on the datasets impede the feasibility of deep learning in edge computing scenarios and suffer from the data scarcity issue. Rather than relying on labeled data, animals learn by interacting with their surroundings and memorizing the relationships between events and objects. This learning paradigm is referred to as associative learning. The successful implementation of associative learning imitates self-learning schemes analogous to animals which resolve the challenges of deep learning. Current state-of-the-art implementations of associative memory are limited to simulations with small-scale and offline paradigms. Thus, this work implements associative memory with an Unmanned Ground Vehicle (UGV) and neuromorphic hardware, specifically Intelâs Loihi, for an online learning scenario. This system emulates the classic associative learning in rats using the UGV in place of the rats. In specific, it successfully reproduces the fear conditioning with no pretraining procedure or labeled datasets. The UGV is rendered capable of autonomously learning the cause-and-effect relationship of the light stimulus and vibration stimulus and exhibiting a movement response to demonstrate the memorization. Hebbian learning dynamics are used to update the synaptic weights during the associative learning process. The Intel Loihi chip is integrated with this online learning system for processing visual signals with a specialized neural assembly. While processing, the Loihiâs average power usages for computing logic and memory are 30 mW and 29 mW, respectively
Adiabatic Quantum Optimization for Associative Memory Recall
Hopfield networks are a variant of associative memory that recall information
stored in the couplings of an Ising model. Stored memories are fixed points for
the network dynamics that correspond to energetic minima of the spin state. We
formulate the recall of memories stored in a Hopfield network using energy
minimization by adiabatic quantum optimization (AQO). Numerical simulations of
the quantum dynamics allow us to quantify the AQO recall accuracy with respect
to the number of stored memories and the noise in the input key. We also
investigate AQO performance with respect to how memories are stored in the
Ising model using different learning rules. Our results indicate that AQO
performance varies strongly with learning rule due to the changes in the energy
landscape. Consequently, learning rules offer indirect methods for
investigating change to the computational complexity of the recall task and the
computational efficiency of AQO.Comment: 22 pages, 11 figures. Updated for clarity and figures, to appear in
Frontiers of Physic
Individual and global adaptation in networks
The structure of complex biological and socio-economic networks affects the selective pressures or behavioural incentives of components in that network, and reflexively, the evolution/behaviour of individuals in those networks changes the structure of such networks over time. Such âadaptive networksâ underlie how gene-regulation networks evolve, how ecological networks self-organise, and how networks of strategic agents co-create social organisations. Although such domains are different in the details, they can each be characterised as networks of self-interested agents where agents alter network connections in the direction that increases their individual utility. Recent work shows that such dynamics are equivalent to associative learning, well-understood in the context of neural networks. Associative learning in neural substrates is the result of mandated learning rules (e.g. Hebbian learning), but in networks of autonomous agents âassociative inductionâ occurs as a result of local individual incentives to alter connections. Using results from a number of recent studies, here we review the theoretical principles that can be transferred between disciplines as a result of this isomorphism, and the implications for the organisation of genetic, social and ecological networks
- âŠ