85 research outputs found
Robust learning algorithms for spiking and rate-based neural networks
Inspired by the remarkable properties of the human brain, the fields of machine learning, computational neuroscience and neuromorphic engineering have achieved significant synergistic progress in the last decade. Powerful neural network models rooted in machine learning have been proposed as models for neuroscience and for applications in neuromorphic engineering. However, the aspect of robustness is often neglected in these models. Both biological and engineered substrates show diverse imperfections that deteriorate the performance of computation models or even prohibit their implementation. This thesis describes three projects aiming at implementing robust learning with local plasticity rules in neural networks. First, we demonstrate the advantages of neuromorphic computations in a pilot study on a prototype chip. Thereby, we quantify the speed and energy consumption of the system compared to a software simulation and show how on-chip learning contributes to the robustness of learning. Second, we present an implementation of spike-based Bayesian inference on accelerated neuromorphic hardware. The model copes, via learning, with the disruptive effects of the imperfect substrate and benefits from the acceleration. Finally, we present a robust model of deep reinforcement learning using local learning rules. It shows how backpropagation combined with neuromodulation could be implemented in a biologically plausible framework. The results contribute to the pursuit of robust and powerful learning networks for biological and neuromorphic substrates
The Impact of Striatal Neuropeptides and Topography on Action Sequence Selection
Many common behaviours are a sequence of several actions. As action sequences are learned their activation often becomes habitual, allowing smooth, rapid, and semi-automatic execution; learning and performing action sequences is central to normal motor function.
The striatum is the primary input nucleus for the basal ganglia and receives glutamatergic cortical afferents. These afferents innervate localised populations of medium spiny neurons (MSNs) and may encode 'action requests'. Striatal interactions ensure that only non-conflicting, high salience requests are selected, but the mechanisms enabling clean, rapid switching between sequential actions are poorly understood.
Substance P (SP) and enkephalin are neuropeptides co-released with GABA by MSNs preferentially expressing D1 or D2 dopamine receptors respectively. SP facilitates subsequent glutamatergic inputs to target MSNs while enkephalin has an inhibitory effect. We construct models of these glutamatergic effects and integrate them into a basal ganglia model to demonstrate that diffuse neuropeptide connectivity enhances action selection. For action sequences with an ordinal structure, patterning SP connectivity to reflect this ordering enhances the selection of correctlyâordered actions and suppresses disordered selection. We also show that selectively pruning SP connections allows contextâsensitive inhibition of specific undesirable requests that otherwise interfere with action group selection.
We then construct a striatal microcircuit model with physical topography and show that inputs to this model generate oscillations in MSN spiking. Input salience and active neuronal density have differentiable impacts on oscillation amplitude and frequency, but the presence of oscillations has little effect on the mean MSN firing rate or action selection.
Our model suggests that neuropeptide interactions enhance the contrast between selected and rejected action requests, and that patterned SP connectivity enhances the selection of ordered sequences. Our model further suggests that striatal topography does not directly impact action selection, but that evoked oscillations may represent an additional form of population coding that could bind together semantically related MSN groups
Recommended from our members
Neural Circuits Underlying Learning and Consolidation
In this work, we develop models of neural circuits and plasticity rules that underlie different forms of learning and memory, with a focus on learning processes that involve multiple brain regions. We begin by surveying the literature on synaptic plasticity rules and implementations of learning algorithms in the brain. Each subsequent chapter presents a model of how a specific aspect of learning is implemented biologically, based on experimental evidence and normative considerations.
We first focus on the neural basis of reinforcement learning in the basal ganglia. We show that in order to enable effective learning when control of behavior is distributed across multiple regions (``off-policy reinforcement learning''), classic models of dopamine activity must be adapted to include an additional action-sensitive component. We also show that the known plasticity rules of direct and indirect-pathway striatal projection neurons are inconsistent with existing models of striatal codes for action.
We propose and find experimental support for a new model of striatal activity driven by efferent input. This model is functionally compatible with striatal plasticity rules and enables simultaneous multiplexing of action-selection and learning signals, a necessary ingredient for off-policy reinforcement learning. We next use an off-policy reinforcement learning model to explain a new experimental finding about the conditions under which learned motor skills are consolidated to be driven by the dorsolateral striatum in rats.
We then shift our focus to consider consolidation more broadly, proposing a general model of the advantages of systems in which memories and learned behaviors are consolidated from short-term to long-term learning pathways. In particular, our model proposes that such architectures enable selective filtering of the set of experiences used for learning, which can be essential in noisy environments with many extraneous stimuli.
In the appendices, we explore other factors relevant to learning algorithms, including the interaction between multiple sensory modalities, and the problem of credit assignment in multi-layer neural networks. In summary, this work presents a varied set of models of different forms of learning in the brain, emphasizing the cooperative role of plasticity rules and multi-regional circuit architecture in producing functionally useful synaptic weight updates
Unveiling the frontiers of deep learning: innovations shaping diverse domains
Deep learning (DL) enables the development of computer models that are
capable of learning, visualizing, optimizing, refining, and predicting data. In
recent years, DL has been applied in a range of fields, including audio-visual
data processing, agriculture, transportation prediction, natural language,
biomedicine, disaster management, bioinformatics, drug design, genomics, face
recognition, and ecology. To explore the current state of deep learning, it is
necessary to investigate the latest developments and applications of deep
learning in these disciplines. However, the literature is lacking in exploring
the applications of deep learning in all potential sectors. This paper thus
extensively investigates the potential applications of deep learning across all
major fields of study as well as the associated benefits and challenges. As
evidenced in the literature, DL exhibits accuracy in prediction and analysis,
makes it a powerful computational tool, and has the ability to articulate
itself and optimize, making it effective in processing data with no prior
training. Given its independence from training data, deep learning necessitates
massive amounts of data for effective analysis and processing, much like data
volume. To handle the challenge of compiling huge amounts of medical,
scientific, healthcare, and environmental data for use in deep learning, gated
architectures like LSTMs and GRUs can be utilized. For multimodal learning,
shared neurons in the neural network for all activities and specialized neurons
for particular tasks are necessary.Comment: 64 pages, 3 figures, 3 table
The role of prefrontal cortex and basal ganglia in model-based and model-free reinforcement learning
Contemporary reinforcement learning (RL) theory suggests that choices can be evaluated either by the model-free (MF) strategy of learning their past worth or the model-based (MB) strategy of predicting their likely consequences based on learning how decision states eventually transition to outcomes. Statistical and computational considerations argue that these strategies should ideally be combined. This thesis aimed to investigate the neural implementation of these two RL strategies and the mechanisms of their interactions. Two non-human primates performed a two-stage decision task designed to elicit and discriminate the use of both MF and MB-RL, while single-neuron activity was recorded from the prefrontal cortex (frontal pole, FP; anterior cingulate cortex, ACC; dorsolateral prefrontal cortex) and striatum (caudate and putamen). Logistic regression analysis revealed that the structure of the task (of MB importance) and the reward history (of MF and MB importance) significantly influenced choice. A trial-by-trial computational analysis also confirmed that choices were made according to a weighted combination of MF and MB- RL, with the influence of the latter approaching 90%. Furthermore, the valuations of both learning methods also influenced response vigour and pupil response. Neural correlates of key elements for MF and MB learning were observed across all brain areas, but functional segregation was also in evidence. Neurons in ACC encoded features of both MF and MB, suggesting a possible role in the arbitration between both strategies. Striatal activity was consistent with a role in value updating by encoding reward prediction errors. Finally, novel neurophysiological evidence was found in favour of the role of the FP in counterfactual processing. In conclusion, this thesis provides insight into the neural implementation of MF and MB-RL computations and their various effects on diverse aspects of behaviour. It supports the parallel operation and integration of the two approaches, while revealing unexpected intricacies
- âŠ