134 research outputs found
Feature Selectivity of the Gamma-Band of the Local Field Potential in Primate Primary Visual Cortex
Extracellular voltage fluctuations (local field potentials, LFPs) reflecting neural mass action are ubiquitous across species and brain regions. Numerous studies have characterized the properties of LFP signals in the cortex to study sensory and motor computations as well as cognitive processes like attention, perception and memory. In addition, its extracranial counterpart – the electroencephalogram – is widely used in clinical applications. However, the link between LFP signals and the underlying activity of local populations of neurons remains largely elusive. Here, we review recent work elucidating the relationship between spiking activity of local neural populations and LFP signals. We focus on oscillations in the gamma-band (30–90 Hz) of the LFP in the primary visual cortex (V1) of the macaque that dominate during visual stimulation. Given that in area V1 much is known about the properties of single neurons and the cortical architecture, it provides an excellent opportunity to study the mechanisms underlying the generation of the LFP
The effect of noise correlations in populations of diversely tuned neurons
The amount of information encoded by networks of neurons critically depends on the correlation structure of their activity. Neurons with similar stimulus preferences tend to have higher noise correlations than others. In homogeneous populations of neurons this limited range correlation structure is highly detrimental to the accuracy of a population code. Therefore, reduced spike count correlations under attention, after adaptation or after learning have been interpreted as evidence for a more efficient population code. Here we analyze the role of limited range correlations in more realistic, heterogeneous population models. We use Fisher information and maximum likelihood decoding to show that reduced correlations do not necessarily improve encoding accuracy. In fact, in populations with more than a few hundred neurons, increasing the level of limited range correlations can substantially improve encoding accuracy. We found that this improvement results from a decrease in noise entropy that is associated with increasing correlations if the marginal distributions are unchanged. Surprisingly, for constant noise entropy and in the limit of large populations the encoding accuracy is independent of both structure and magnitude of noise correlations
Optimal Population Coding, Revisited
Cortical circuits perform the computations underlying rapid perceptual decisions within a few dozen milliseconds with each neuron emitting only a few spikes. Under these conditions, the theoretical analysis of neural population codes is challenging, as the most commonly used theoretical tool – Fisher information – can lead to erroneous conclusions about the optimality of different coding schemes. Here we revisit the effect of tuning function width and correlation structure on neural population codes based on ideal observer analysis in both a discrimination and reconstruction task. We show that the optimal tuning function width and the optimal correlation structure in both paradigms strongly depend on the available decoding time in a very similar way. In contrast, population codes optimized for Fisher information do not depend on decoding time and are severely suboptimal when only few spikes are available. In addition, we use the neurometric functions of the ideal observer in the classification task to investigate the differential coding properties of these Fisher-optimal codes for fine and coarse discrimination. We find that the discrimination error for these codes does not decrease to zero with increasing population size, even in simple coarse discrimination tasks. Our results suggest that quite different population codes may be optimal for rapid decoding in cortical computations than those inferred from the optimization of Fisher information
Comparing the Feature Selectivity of the Gamma-Band of the Local Field Potential and the Underlying Spiking Activity in Primate Visual Cortex
The local field potential (LFP), comprised of low-frequency extra-cellular voltage fluctuations, has been used extensively to study the mechanisms of brain function. In particular, oscillations in the gamma-band (30–90 Hz) are ubiquitous in the cortex of many species during various cognitive processes. Surprisingly little is known about the underlying biophysical processes generating this signal. Here, we examine the relationship of the local field potential to the activity of localized populations of neurons by simultaneously recording spiking activity and LFP from the primary visual cortex (V1) of awake, behaving macaques. The spatial organization of orientation tuning and ocular dominance in this area provides an excellent opportunity to study this question, because orientation tuning is organized at a scale around one order of magnitude finer than the size of ocular dominance columns. While we find a surprisingly weak correlation between the preferred orientation of multi-unit activity and gamma-band LFP recorded on the same tetrode, there is a strong correlation between the ocular preferences of both signals. Given the spatial arrangement of orientation tuning and ocular dominance, this leads us to conclude that the gamma-band of the LFP seems to sample an area considerably larger than orientation columns. Rather, its spatial resolution lies at the scale of ocular dominance columns
DataJoint: managing big scientific data using MATLAB or Python
The rise of big data in modern research poses serious challenges for data management: Large and intricate datasets from diverse instrumentation must be precisely aligned, annotated, and processed in a variety of ways to extract new insights. While high levels of data integrity are expected, research teams have diverse backgrounds, are geographically dispersed, and rarely possess a primary interest in data science. Here we describe DataJoint, an open-source toolbox designed for manipulating and processing scientific data under the relational data model. Designed for scientists who need a flexible and expressive database language with few basic concepts and operations, DataJoint facilitates multi-user access, efficient queries, and distributed computing. With implementations in both MATLAB and Python, DataJoint is not limited to particular file formats, acquisition systems, or data modalities and can be quickly adapted to new experimental designs. DataJoint and related resources are available at http://datajoint.github.com
Understanding Robustness and Generalization of Artificial Neural Networks Through Fourier Masks
Despite the enormous success of artificial neural networks (ANNs) in many disciplines, the characterization of their computations and the origin of key properties such as generalization and robustness remain open questions. Recent literature suggests that robust networks with good generalization properties tend to be biased toward processing low frequencies in images. To explore the frequency bias hypothesis further, we develop an algorithm that allows us to learn modulatory masks highlighting the essential input frequencies needed for preserving a trained network's performance. We achieve this by imposing invariance in the loss with respect to such modulations in the input frequencies. We first use our method to test the low-frequency preference hypothesis of adversarially trained or data-augmented networks. Our results suggest that adversarially robust networks indeed exhibit a low-frequency bias but we find this bias is also dependent on directions in frequency space. However, this is not necessarily true for other types of data augmentation. Our results also indicate that the essential frequencies in question are effectively the ones used to achieve generalization in the first place. Surprisingly, images seen through these modulatory masks are not recognizable and resemble texture-like patterns
Few-Shot Attribute Learning
Semantic concepts are frequently defined by combinations of underlying
attributes. As mappings from attributes to classes are often simple,
attribute-based representations facilitate novel concept learning with zero or
few examples. A significant limitation of existing attribute-based learning
paradigms, such as zero-shot learning, is that the attributes are assumed to be
known and fixed. In this work we study the rapid learning of attributes that
were not previously labeled. Compared to standard few-shot learning of semantic
classes, in which novel classes may be defined by attributes that were relevant
at training time, learning new attributes imposes a stiffer challenge. We found
that supervised learning with training attributes does not generalize well to
new test attributes, whereas self-supervised pre-training brings significant
improvement. We further experimented with random splits of the attribute space
and found that predictability of test attributes provides an informative
estimate of a model's generalization ability.Comment: Technical report, 25 page
Recommended from our members
Brain-inspired replay for continual learning with artificial neural networks.
Funder: International Brain Research Organization (IBRO); doi: https://doi.org/10.13039/501100001675Artificial neural networks suffer from catastrophic forgetting. Unlike humans, when these networks are trained on something new, they rapidly forget what was learned before. In the brain, a mechanism thought to be important for protecting memories is the reactivation of neuronal activity patterns representing those memories. In artificial neural networks, such memory replay can be implemented as 'generative replay', which can successfully - and surprisingly efficiently - prevent catastrophic forgetting on toy examples even in a class-incremental learning scenario. However, scaling up generative replay to complicated problems with many tasks or complex inputs is challenging. We propose a new, brain-inspired variant of replay in which internal or hidden representations are replayed that are generated by the network's own, context-modulated feedback connections. Our method achieves state-of-the-art performance on challenging continual learning benchmarks (e.g., class-incremental learning on CIFAR-100) without storing data, and it provides a novel model for replay in the brain
- …