9,947 research outputs found
Multiscale relevance and informative encoding in neuronal spike trains
Neuronal responses to complex stimuli and tasks can encompass a wide range of
time scales. Understanding these responses requires measures that characterize
how the information on these response patterns are represented across multiple
temporal resolutions. In this paper we propose a metric -- which we call
multiscale relevance (MSR) -- to capture the dynamical variability of the
activity of single neurons across different time scales. The MSR is a
non-parametric, fully featureless indicator in that it uses only the time
stamps of the firing activity without resorting to any a priori covariate or
invoking any specific structure in the tuning curve for neural activity. When
applied to neural data from the mEC and from the ADn and PoS regions of
freely-behaving rodents, we found that neurons having low MSR tend to have low
mutual information and low firing sparsity across the correlates that are
believed to be encoded by the region of the brain where the recordings were
made. In addition, neurons with high MSR contain significant information on
spatial navigation and allow to decode spatial position or head direction as
efficiently as those neurons whose firing activity has high mutual information
with the covariate to be decoded and significantly better than the set of
neurons with high local variations in their interspike intervals. Given these
results, we propose that the MSR can be used as a measure to rank and select
neurons for their information content without the need to appeal to any a
priori covariate.Comment: 38 pages, 16 figure
Deep Neural Networks Rival the Representation of Primate IT Cortex for Core Visual Object Recognition
The primate visual system achieves remarkable visual object recognition
performance even in brief presentations and under changes to object exemplar,
geometric transformations, and background variation (a.k.a. core visual object
recognition). This remarkable performance is mediated by the representation
formed in inferior temporal (IT) cortex. In parallel, recent advances in
machine learning have led to ever higher performing models of object
recognition using artificial deep neural networks (DNNs). It remains unclear,
however, whether the representational performance of DNNs rivals that of the
brain. To accurately produce such a comparison, a major difficulty has been a
unifying metric that accounts for experimental limitations such as the amount
of noise, the number of neural recording sites, and the number trials, and
computational limitations such as the complexity of the decoding classifier and
the number of classifier training examples. In this work we perform a direct
comparison that corrects for these experimental limitations and computational
considerations. As part of our methodology, we propose an extension of "kernel
analysis" that measures the generalization accuracy as a function of
representational complexity. Our evaluations show that, unlike previous
bio-inspired models, the latest DNNs rival the representational performance of
IT cortex on this visual object recognition task. Furthermore, we show that
models that perform well on measures of representational performance also
perform well on measures of representational similarity to IT and on measures
of predicting individual IT multi-unit responses. Whether these DNNs rely on
computational mechanisms similar to the primate visual system is yet to be
determined, but, unlike all previous bio-inspired models, that possibility
cannot be ruled out merely on representational performance grounds.Comment: 35 pages, 12 figures, extends and expands upon arXiv:1301.353
- …