1,510 research outputs found
Population coding in sparsely connected networks of noisy neurons
This study examines the relationship between population coding and spatial connection statistics in networks of noisy neurons. Encoding of sensory information in the neocortex is thought to require coordinated neural populations, because individual cortical neurons respond to a wide range of stimuli, and exhibit highly variable spiking in response to repeated stimuli. Population coding is rooted in network structure, because cortical neurons receive information only from other neurons, and because the information they encode must be decoded by other neurons, if it is to affect behavior. However, population coding theory has often ignored network structure, or assumed discrete, fully connected populations (in contrast with the sparsely connected, continuous sheet of the cortex). In this study, we modeled a sheet of cortical neurons with sparse, primarily local connections, and found that a network with this structure could encode multiple internal state variables with high signal-to-noise ratio. However, we were unable to create high-fidelity networks by instantiating connections at random according to spatial connection probabilities. In our models, high-fidelity networks required additional structure, with higher cluster factors and correlations between the inputs to nearby neurons
Noise-enhanced computation in a model of a cortical column
Varied sensory systems use noise in order to enhance detection of weak
signals. It has been conjectured in the literature that this effect, known as
stochastic resonance, may take place in central cognitive processes such as the
memory retrieval of arithmetical multiplication. We show in a simplified model
of cortical tissue, that complex arithmetical calculations can be carried out
and are enhanced in the presence of a stochastic background. The performance is
shown to be positively correlated to the susceptibility of the network, defined
as its sensitivity to a variation of the mean of its inputs. For nontrivial
arithmetic tasks such as multiplication, stochastic resonance is an emergent
property of the microcircuitry of the model network
Inferring Synaptic Structure in presence of Neural Interaction Time Scales
Biological networks display a variety of activity patterns reflecting a web
of interactions that is complex both in space and time. Yet inference methods
have mainly focused on reconstructing, from the network's activity, the spatial
structure, by assuming equilibrium conditions or, more recently, a
probabilistic dynamics with a single arbitrary time-step. Here we show that,
under this latter assumption, the inference procedure fails to reconstruct the
synaptic matrix of a network of integrate-and-fire neurons when the chosen time
scale of interaction does not closely match the synaptic delay or when no
single time scale for the interaction can be identified; such failure,
moreover, exposes a distinctive bias of the inference method that can lead to
infer as inhibitory the excitatory synapses with interaction time scales longer
than the model's time-step. We therefore introduce a new two-step method, that
first infers through cross-correlation profiles the delay-structure of the
network and then reconstructs the synaptic matrix, and successfully test it on
networks with different topologies and in different activity regimes. Although
step one is able to accurately recover the delay-structure of the network, thus
getting rid of any \textit{a priori} guess about the time scales of the
interaction, the inference method introduces nonetheless an arbitrary time
scale, the time-bin used to binarize the spike trains. We therefore
analytically and numerically study how the choice of affects the inference
in our network model, finding that the relationship between the inferred
couplings and the real synaptic efficacies, albeit being quadratic in both
cases, depends critically on for the excitatory synapses only, whilst
being basically independent of it for the inhibitory ones
Toward a Robust Sparse Data Representation for Wireless Sensor Networks
Compressive sensing has been successfully used for optimized operations in
wireless sensor networks. However, raw data collected by sensors may be neither
originally sparse nor easily transformed into a sparse data representation.
This paper addresses the problem of transforming source data collected by
sensor nodes into a sparse representation with a few nonzero elements. Our
contributions that address three major issues include: 1) an effective method
that extracts population sparsity of the data, 2) a sparsity ratio guarantee
scheme, and 3) a customized learning algorithm of the sparsifying dictionary.
We introduce an unsupervised neural network to extract an intrinsic sparse
coding of the data. The sparse codes are generated at the activation of the
hidden layer using a sparsity nomination constraint and a shrinking mechanism.
Our analysis using real data samples shows that the proposed method outperforms
conventional sparsity-inducing methods.Comment: 8 page
Mechanisms of Zero-Lag Synchronization in Cortical Motifs
Zero-lag synchronization between distant cortical areas has been observed in
a diversity of experimental data sets and between many different regions of the
brain. Several computational mechanisms have been proposed to account for such
isochronous synchronization in the presence of long conduction delays: Of
these, the phenomenon of "dynamical relaying" - a mechanism that relies on a
specific network motif - has proven to be the most robust with respect to
parameter mismatch and system noise. Surprisingly, despite a contrary belief in
the community, the common driving motif is an unreliable means of establishing
zero-lag synchrony. Although dynamical relaying has been validated in empirical
and computational studies, the deeper dynamical mechanisms and comparison to
dynamics on other motifs is lacking. By systematically comparing
synchronization on a variety of small motifs, we establish that the presence of
a single reciprocally connected pair - a "resonance pair" - plays a crucial
role in disambiguating those motifs that foster zero-lag synchrony in the
presence of conduction delays (such as dynamical relaying) from those that do
not (such as the common driving triad). Remarkably, minor structural changes to
the common driving motif that incorporate a reciprocal pair recover robust
zero-lag synchrony. The findings are observed in computational models of
spiking neurons, populations of spiking neurons and neural mass models, and
arise whether the oscillatory systems are periodic, chaotic, noise-free or
driven by stochastic inputs. The influence of the resonance pair is also robust
to parameter mismatch and asymmetrical time delays amongst the elements of the
motif. We call this manner of facilitating zero-lag synchrony resonance-induced
synchronization, outline the conditions for its occurrence, and propose that it
may be a general mechanism to promote zero-lag synchrony in the brain.Comment: 41 pages, 12 figures, and 11 supplementary figure
- …