13 research outputs found
Recommended from our members
Synaptic plasticity and memory addressing in biological and artificial neural networks
Biological brains are composed of neurons, interconnected by synapses to create large complex networks. Learning and memory occur, in large part, due to synaptic plasticity -- modifications in the efficacy of information transmission through these synaptic connections. Artificial neural networks model these with neural "units" which communicate through synaptic weights. Models of learning and memory propose synaptic plasticity rules that describe and predict the weight modifications. An equally important but under-evaluated question is the selection of \textit{which} synapses should be updated in response to a memory event. In this work, we attempt to separate the questions of synaptic plasticity from that of memory addressing.
Chapter 1 provides an overview of the problem of memory addressing and a summary of the solutions that have been considered in computational neuroscience and artificial intelligence, as well as those that may exist in biology. Chapter 2 presents in detail a solution to memory addressing and synaptic plasticity in the context of familiarity detection, suggesting strong feedforward weights and anti-Hebbian plasticity as the respective mechanisms. Chapter 3 proposes a model of recall, with storage performed by addressing through local third factors and neo-Hebbian plasticity, and retrieval by content-based addressing. In Chapter 4, we consider the problem of concurrent memory consolidation and memorization. Both storage and retrieval are performed by content-based addressing, but the plasticity rule itself is implemented by gradient descent, modulated according to whether an item should be stored in a distributed manner or memorized verbatim. However, the classical method for computing gradients in recurrent neural networks, backpropagation through time, is generally considered unbiological. In Chapter 5 we suggest a more realistic implementation through an approximation of recurrent backpropagation.
Taken together, these results propose a number of potential mechanisms for memory storage and retrieval, each of which separates the mechanism of synaptic updating -- plasticity -- from that of synapse selection -- addressing. Explicit studies of memory addressing may find applications not only in artificial intelligence but also in biology. In artificial networks, for example, selectively updating memories in large language models can help improve user privacy and security. In biological ones, understanding memory addressing can help with health outcomes and treating memory-based illnesses such as Alzheimers or PTSD
Recommended from our members
Meta-learning Hebbian plasticity for continual familiarity detection
Memories are stored and recalled throughout the lifetime of an animal, but many models of memory, including previous models of familiarity detection, do not operate in a continuous manner. We consider a family of models that recognize previously experienced stimuli and, importantly, operate and learn continuously. Specifically, we investigate a learning paradigm in which stimuli are presented in a streaming fashion with repetitions at various intervals, and the subject/model must report whether the current stimulus has previously appeared in the stream. We propose a feedforward network architecture with ongoing plasticity in the synaptic weight matrix. Parameters governing plasticity and static network parameters are meta-learned using gradient descent to optimize the continual familiarity detection process. This architecture, unlike recurrent networks without ongoing plasticity, generalizes easily over a range of repeat intervals even if trained with a single interval. We show that an anti-Hebbian plasticity rule (co-activated neurons cause synaptic depression) enables repeat detection over much longer intervals than a Hebbian one, and this is the solution most readily found by meta-learning. This rule leads to experimentally observed features such as repeat suppression in the hidden layer neurons. In contrast to previous theoretical work, the capacity of these networks remains constant across their lifetimes, meaning that pairs of stimuli with a given temporal separation are stored and recognized as familiar independent of the network's input history. We also consider learning rules that use an external gating circuit to control plasticity. Collectively, these models demonstrate a range of different psychometric curves that we compare to human performance.
Keywords: learning, memory, recognition, familiarity, novelty detection, meta-learning, Hebbian, synaptic plasticit
Recommended from our members
Biological learning in key-value memory networks
In neuroscience, classical Hopfield networks are the standard biologically plausible model of long-term memory, relying on Hebbian plasticity for storage and attractor dynamics for recall. In contrast, memory-augmented neural networks in machine learning commonly use a key-value mechanism to store and read out memories in a single step. Such augmented networks achieve impressive feats of memory compared to traditional variants, yet it remains unclear whether they can be implemented by biological systems. In our work, we bridge this gap by proposing a set of of biologically plausible three-factor plasticity rules for a basic feedforward key-value memory network. Keys are stored in the input-to-hidden synaptic weights by a "non-Hebbian" rule, controlled only by pre-synaptic activity, and modulated by local third factors which represent dendtritic spikes. Values are stored in the hidden-to-output weights by a Hebbian rule, with the pre-synaptic neuron selected through softmax attention which represents recurrent inhibition. The same rules are recovered when network parameters are meta-learned. Our network performs on par with classical Hopfield networks on autoassociative memory tasks and can be naturally extended to correlated inputs, continual recall, heteroassociative memory, and sequence learning. Importantly, since memories are stored in slots indexed by hidden layer neurons, unlike the fully distributed representation in the classical Hopfield network, they can be individually selected for extended storage or rapid decay. Finally, our memory network can easily be incorporated into a larger neural system, either as a memory bank for an external controller, or as a fast learning system used in conjunction with a slow one. Overall, our results suggest a compelling alternative to the classical Hopfield network as a model of biological long-term memory.
Keywords: learning, memory, synaptic plasticity, Hebbian, key-value memory, neural network, three-factor plasticit
Time-domain diffuse correlation spectroscopy
Physiological monitoring of oxygen delivery to the brain has great significance for improving the management of patients at risk for brain injury. Diffuse correlation spectroscopy (DCS) is a rapidly growing optical technology able to non-invasively assess the blood flow index (BFi) at the bedside. The current limitations of DCS are the contamination introduced by extracerebral tissue and the need to know the tissue's optical properties to correctly quantify the BFi. To overcome these limitations, we have developed a new technology for time-resolved diffuse correlation spectroscopy. By operating DCS in the time domain (TD-DCS), we are able to simultaneously acquire the temporal point-spread function to quantify tissue optical properties and the autocorrelation function to quantify the BFi. More importantly, by applying time-gated strategies to the DCS autocorrelation functions, we are able to differentiate between short and long photon paths through the tissue and determine the BFi for different depths. Here, we present the novel device and we report the first experiments in tissue-like phantoms and in rodents. The TD-DCS method opens many possibilities for improved non-invasive monitoring of oxygen delivery in humans
Time-domain diffuse correlation spectroscopy: instrument prototype, preliminary measurements, and theoretical modeling
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2017.This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.Cataloged from student-submitted PDF version of thesis.Includes bibliographical references (pages 86-91).Near-infrared spectroscopy (NIRS) is an emerging diffuse optical imaging tool with both clinical and academic applications such as functional brain imaging, breast cancer detection, and cerebral health monitoring. Due to its non-invasiveness, high spatial and temporal resolution, and portability, it has been rapidly growing in popularity over the last 40 years. The technique relies on near-infrared light to measure optical properties { scattering and absorption { which can then be used to infer details of the underlying tissue physiology. Diffuse correlation spectroscopy (DCS) is a complimentary optical technique that relies on long-coherence laser light, also in the near-infrared range, to measure dynamical properties of a medium { in the biomedical context, blood ow. While NIRS and DCS can be used in conjunction to provide even more powerful information, they require separate instrumentation, resulting in reduced portability and difficulty in bedside monitoring. In brain imaging applications, both NIRS and DCS suer from confounds due to layers surrounding the brain, such as the scalp and skull. While this issue has been addressed in NIRS using time-resolved instrumentation known as time-domain (TD) NIRS, it has been largely ignored in the context of DCS. In this work, we demonstrate a novel time-domain diffuse correlation spectroscopy (TD-DCS) technique embodied in a single instrument capable of simultaneously measuring optical and dynamical properties. Along with maintaining portability, the instrument reduces error by directly measuring the absorption and scattering values necessary for precise ow estimation, and removes a major confounding factor by suppressing unwanted signal from superficial layers through time-gating. We describe the construction of the first instrument prototype and demonstrate the depth resolution proof-of-concept with measurements of multi-layer media. We further discuss the theoretical considerations of modeling the light interaction with tissue, necessary for reliable estimates.by Danil Tyulmankov.M. Eng
Periodic flashing coordinated reset stimulation paradigm reduces sensitivity to ON and OFF period durations.
Pathological synchronization in the basal ganglia network has been considered an important component of Parkinson's disease pathophysiology. An established treatment for some patients with Parkinson's disease is deep brain stimulation, in which a tonic high-frequency pulse train is delivered to target regions of the brain. In recent years, a novel neuromodulation paradigm called coordinated reset stimulation has been proposed, which aims to reverse the pathological synchrony by sequentially delivering short high-frequency bursts to distinct sub-regions of the pathologically synchronized network, with an average intra-burst interval for each sub-region corresponding to period of the pathological oscillation. It has further been proposed that the resultant desynchronization can be enhanced when stimulation is interrupted periodically, and that it is particularly beneficial to precisely tune the stimulation ON and OFF time-windows to the underlying pathological frequency. Pre-clinical and clinical studies of coordinated reset stimulation have relied on these proposals for their stimulation protocols. In this study, we present a modified ON-OFF coordinated reset stimulation paradigm called periodic flashing and study its behavior through computational modeling using the Kuramoto coupled phase oscillator model. We demonstrate that in contrast to conventional coordinated reset stimulation, the periodic flashing variation does not exhibit a need for precise turning of the ON-OFF periods to the pathological frequency, and demonstrates desynchronization for a wide range of ON and OFF periods. We provide a mechanistic explanation for the previously observed sensitivities and demonstrate that they are an artifact of the specific ON-OFF cycling paradigm used. As a practical consequence, the periodic flashing paradigm simplifies the tuning of optimal stimulation parameters by decreasing the dimension of the search space. It also suggests new, more flexible ways of delivering coordinated reset stimulation
Memory consolidation through combined burst-induced homeostatic reset and structural plasticity
editorial reviewedNeurons adapt their connections with each other through synaptic plasticity, driven by correlations in their spiking activity (Fig. 1A). Additionally, neuronal networks undergo global changes in rhythmic activity that correspond to different brain states, defined by switches in neuronal activity and orchestrated by neuromodulators. A well-known example is the transition from a state of learning during active waking to rest during quiet waking, which corresponds to a switch in neuronal activity from tonic firing to bursting (Fig. 1B). This raises the question of how switching from tonic firing to bursting affects the outcome of synaptic plasticity and whether it can support memory consolidation.
Recently, we have shown for a variety of synaptic plasticity models that bursting leads to a homeostatic reset, in which synaptic efficacy returns to a fixed baseline value irrespective of the starting point. This homeostatic reset causes the network to forget any learned information [1]. To address this issue, we propose an additional structural plasticity mechanism in which short-term changes in synaptic efficacy – evolving according to traditional plasticity rules – drive long-lasting morphological changes such as spine growth or insertion of new AMPA receptors. While synaptic efficacy undergoes homeostatic reset during bursting, information is consolidated through structural plasticity on a longer timescale.
We demonstrate the utility of this mechanism in a network of neurons using a conductance-based neuronal model that can switch from tonic firing to bursting along with a calcium-based synaptic rule to drive changes in synaptic efficacy. We investigate three regimes of switches in neuronal activity and plasticity mechanisms, denoted S1, S2, S3 (Fig. 1C). In S1, as a control condition, tonic firing is interleaved with periods of neuronal inactivity – mimicking bursting blockers – and a traditional plasticity rule, while in S2, tonic firing is separated by periods of bursting, leading to homeostatic reset in synaptic efficacy. Configuration S3 is identical as S2 but also includes our proposed burst-driven structural plasticity.
In our first memory task (Fig. 1D), we show that the signal-to-noise (SNR) is improved over repeated switches only in S3. In a simple pattern recognition task (Fig. 1E), blocking bursting activity (S1) makes the network fragile to noise, and blocking structural plasticity during bursting leads to complete forgetting (S2), neither of which occurs in S3. Finally, in a MNIST recognition task (Fig. 1F), we confirm that memory consolidation occurs with S3 by showing a stronger receptive field that consolidates during switches from tonic to burst and is robust to noise.
In this work, we shed light on the under-investigated role of switches in neuronal firing patterns for synaptic plasticity. Traditional plasticity rules result in a burst-induced homeostatic reset of synaptic efficacy, which is incompatible with memory consolidation. Our burst-driven structural plasticity proposes a solution to this problem, bridging the gap between switches in tonic firing to bursting, learning, and memory consolidation, and suggesting new ways to improve machine learning algorithms.
References
[1] Jacquerie K, Minne C, Ponnet J, et al. Switches to slow rhythmic neuronal activity lead to a plasticity-induced reset in synaptic weights. preprint, Biorxiv (2022)
Switching from tonic firing to bursting: implications on learning and memory
editorial reviewedThe brain's ability to learn and remember information relies on synaptic plasticity, the process by which neurons change the strength of their connections in response to their spiking activity. In parallel, the switch between different brain states, such as from active to quiet wakefulness, involves a transition in neuronal activity from tonic firing to bursting. It raises questions about how switches between different states affect synaptic plasticity and memory consolidation.
Recent research has revealed that bursting leads to a homeostatic reset of synaptic efficacy, where synaptic efficacy returns to a baseline value regardless of its starting point, causing the network to forget any learned information. To address this issue, we propose a new mechanism called burst-driven structural plasticity, that combines early changes in synaptic efficacy with long-lasting morphological changes such as spine growth or new protein synthesis.
Using a conductance-based neuronal model with a calcium-based plasticity rule, we demonstrate the utility of the proposed mechanism in a network that learns to recognize hand-written digits from the MNIST dataset. Our results show that the combination of switching from tonic firing to bursting with structural plasticity improves memory consolidation and enhances network robustness to noise. Conversely, blocking bursting and its burst-driven structural plasticity leads to forgetting.
In conclusion, this study highlights the importance of firing patterns in synaptic plasticity and proposes a solution to bridge the gap between the switches from tonic firing to bursting, learning and memory consolidation
Time domain diffuse correlation spectroscopy: modeling the effects of laser coherence length and instrument response function
Diffuse correlation spectroscopy (DCS) is an optical technique that non-invasively quantifies an index of blood flow (BFi) by measuring the temporal autocorrelation function of the intensity fluctuations of light diffusely remitted from the tissue. Traditional DCS measurements use continuous wave (CW) lasers with coherence lengths longer than the photon path lengths in the sample to ensure that the diffusely remitted light is coherent and generates a speckle pattern. Recently, we proposed time domain DCS (TD-DCS) to allow measurements of the speckle fluctuations for specific path lengths of light through the tissue, which has the distinct advantage of permitting an analysis of selected long path lengths of light to improve the depth sensitivity of the measurement. However, compared to CW-DCS, factors including the instrument response function (IRF), the detection gate width, and the finite coherence length need to be considered in the model analysis of the experimental data. Here we present a TD-DCS model describing how the intensity autocorrelation functions measured for different path lengths of light depend on the coherence length, pulse width of the laser, detection gate width, IRF, BFi, and optical properties of the scattering sample. Predictions of the model are compared with experimental results using a homogeneous liquid phantom sample that mimics human tissue optical properties. The BFis obtained from the TD-DCS model for different path lengths of light agree with the BFi obtained from CW-DCS measurements, while the standard simplified model underestimates the BFi by a factor of ∼2. This Letter establishes the theoretical foundation of the TD-DCS technique and provides guidance for future BFi measurements in tissue.National Institutes of Health (NIH) (R01EB025145, R01-MH111359 (DAB), R01-NS100750 (SAC), R21NS094828 (MAF)); Office of the Secretary of Defense (OSD); U.S. Air Force (USAF) (FA8721-05-C-0002, FA8702-15-D-0001). (R01EB025145 - National Institutes of Health (NIH); R01-MH111359 - National Institutes of Health (NIH); R01-NS100750 - National Institutes of Health (NIH); R21NS094828 - National Institutes of Health (NIH); Office of the Secretary of Defense (OSD); FA8721-05-C-0002 - U.S. Air Force (USAF); FA8702-15-D-0001 - U.S. Air Force (USAF))https://www.osapublishing.org/ol/abstract.cfm?uri=ol-43-12-2756Accepted manuscrip