8,747 research outputs found
State Dependence of Stimulus-Induced Variability Tuning in Macaque MT
Behavioral states marked by varying levels of arousal and attention modulate
some properties of cortical responses (e.g. average firing rates or pairwise
correlations), yet it is not fully understood what drives these response
changes and how they might affect downstream stimulus decoding. Here we show
that changes in state modulate the tuning of response variance-to-mean ratios
(Fano factors) in a fashion that is neither predicted by a Poisson spiking
model nor changes in the mean firing rate, with a substantial effect on
stimulus discriminability. We recorded motion-sensitive neurons in middle
temporal cortex (MT) in two states: alert fixation and light, opioid
anesthesia. Anesthesia tended to lower average spike counts, without decreasing
trial-to-trial variability compared to the alert state. Under anesthesia,
within-trial fluctuations in excitability were correlated over longer time
scales compared to the alert state, creating supra-Poisson Fano factors. In
contrast, alert-state MT neurons have higher mean firing rates and largely
sub-Poisson variability that is stimulus-dependent and cannot be explained by
firing rate differences alone. The absence of such stimulus-induced variability
tuning in the anesthetized state suggests different sources of variability
between states. A simple model explains state-dependent shifts in the
distribution of observed Fano factors via a suppression in the variance of gain
fluctuations in the alert state. A population model with stimulus-induced
variability tuning and behaviorally constrained information-limiting
correlations explores the potential enhancement in stimulus discriminability by
the cortical population in the alert state.Comment: 36 pages, 18 figure
Signatures of criticality arise in simple neural population models with correlations
Large-scale recordings of neuronal activity make it possible to gain insights
into the collective activity of neural ensembles. It has been hypothesized that
neural populations might be optimized to operate at a 'thermodynamic critical
point', and that this property has implications for information processing.
Support for this notion has come from a series of studies which identified
statistical signatures of criticality in the ensemble activity of retinal
ganglion cells. What are the underlying mechanisms that give rise to these
observations? Here we show that signatures of criticality arise even in simple
feed-forward models of retinal population activity. In particular, they occur
whenever neural population data exhibits correlations, and is randomly
sub-sampled during data analysis. These results show that signatures of
criticality are not necessarily indicative of an optimized coding strategy, and
challenge the utility of analysis approaches based on equilibrium
thermodynamics for understanding partially observed biological systems.Comment: 36 pages, LaTeX; added journal reference on page 1, added link to
code repositor
Neural population coding: combining insights from microscopic and mass signals
Behavior relies on the distributed and coordinated activity of neural populations. Population activity can be measured using multi-neuron recordings and neuroimaging. Neural recordings reveal how the heterogeneity, sparseness, timing, and correlation of population activity shape information processing in local networks, whereas neuroimaging shows how long-range coupling and brain states impact on local activity and perception. To obtain an integrated perspective on neural information processing we need to combine knowledge from both levels of investigation. We review recent progress of how neural recordings, neuroimaging, and computational approaches begin to elucidate how interactions between local neural population activity and large-scale dynamics shape the structure and coding capacity of local information representations, make them state-dependent, and control distributed populations that collectively shape behavior
Optimal Population Coding, Revisited
Cortical circuits perform the computations underlying rapid perceptual decisions within a few dozen milliseconds with each neuron emitting only a few spikes. Under these conditions, the theoretical analysis of neural population codes is challenging, as the most commonly used theoretical tool – Fisher information – can lead to erroneous conclusions about the optimality of different coding schemes. Here we revisit the effect of tuning function width and correlation structure on neural population codes based on ideal observer analysis in both a discrimination and reconstruction task. We show that the optimal tuning function width and the optimal correlation structure in both paradigms strongly depend on the available decoding time in a very similar way. In contrast, population codes optimized for Fisher information do not depend on decoding time and are severely suboptimal when only few spikes are available. In addition, we use the neurometric functions of the ideal observer in the classification task to investigate the differential coding properties of these Fisher-optimal codes for fine and coarse discrimination. We find that the discrimination error for these codes does not decrease to zero with increasing population size, even in simple coarse discrimination tasks. Our results suggest that quite different population codes may be optimal for rapid decoding in cortical computations than those inferred from the optimization of Fisher information
When do correlations increase with firing rates in recurrent networks?
A central question in neuroscience is to understand how noisy firing patterns are used to transmit information. Because neural spiking is noisy, spiking patterns are often quantified via pairwise correlations, or the probability that two cells will spike coincidentally, above and beyond their baseline firing rate. One observation frequently made in experiments, is that correlations can increase systematically with firing rate. Theoretical studies have determined that stimulus-dependent correlations that increase with firing rate can have beneficial effects on information coding; however, we still have an incomplete understanding of what circuit mechanisms do, or do not, produce this correlation-firing rate relationship. Here, we studied the relationship between pairwise correlations and firing rates in recurrently coupled excitatory-inhibitory spiking networks with conductance-based synapses. We found that with stronger excitatory coupling, a positive relationship emerged between pairwise correlations and firing rates. To explain these findings, we used linear response theory to predict the full correlation matrix and to decompose correlations in terms of graph motifs. We then used this decomposition to explain why covariation of correlations with firing rate—a relationship previously explained in feedforward networks driven by correlated input—emerges in some recurrent networks but not in others. Furthermore, when correlations covary with firing rate, this relationship is reflected in low-rank structure in the correlation matrix
- …