51 research outputs found
Modular Acquisition and Stimulation System for Timestamp-Driven Neuroscience Experiments
Dedicated systems are fundamental for neuroscience experimental protocols
that require timing determinism and synchronous stimuli generation. We
developed a data acquisition and stimuli generator system for neuroscience
research, optimized for recording timestamps from up to 6 spiking neurons and
entirely specified in a high-level Hardware Description Language (HDL). Despite
the logic complexity penalty of synthesizing from such a language, it was
possible to implement our design in a low-cost small reconfigurable device.
Under a modular framework, we explored two different memory arbitration schemes
for our system, evaluating both their logic element usage and resilience to
input activity bursts. One of them was designed with a decoupled and latency
insensitive approach, allowing for easier code reuse, while the other adopted a
centralized scheme, constructed specifically for our application. The usage of
a high-level HDL allowed straightforward and stepwise code modifications to
transform one architecture into the other. The achieved modularity is very
useful for rapidly prototyping novel electronic instrumentation systems
tailored to scientific research.Comment: Preprint submitted to ARC 2015. Extended: 16 pages, 10 figures. The
final publication is available at link.springer.co
Entropy and information in neural spike trains: Progress on the sampling problem
The major problem in information theoretic analysis of neural responses and
other biological data is the reliable estimation of entropy--like quantities
from small samples. We apply a recently introduced Bayesian entropy estimator
to synthetic data inspired by experiments, and to real experimental spike
trains. The estimator performs admirably even very deep in the undersampled
regime, where other techniques fail. This opens new possibilities for the
information theoretic analysis of experiments, and may be of general interest
as an example of learning from limited data.Comment: 7 pages, 4 figures; referee suggested changes, accepted versio
Reconstruction of metabolic networks from high-throughput metabolite profiling data: in silico analysis of red blood cell metabolism
We investigate the ability of algorithms developed for reverse engineering of
transcriptional regulatory networks to reconstruct metabolic networks from
high-throughput metabolite profiling data. For this, we generate synthetic
metabolic profiles for benchmarking purposes based on a well-established model
for red blood cell metabolism. A variety of data sets is generated, accounting
for different properties of real metabolic networks, such as experimental
noise, metabolite correlations, and temporal dynamics. These data sets are made
available online. We apply ARACNE, a mainstream transcriptional networks
reverse engineering algorithm, to these data sets and observe performance
comparable to that obtained in the transcriptional domain, for which the
algorithm was originally designed.Comment: 14 pages, 3 figures. Presented at the DIMACS Workshop on Dialogue on
Reverse Engineering Assessment and Methods (DREAM), Sep 200
A machine learning pipeline for discriminant pathways identification
Motivation: Identifying the molecular pathways more prone to disruption
during a pathological process is a key task in network medicine and, more in
general, in systems biology.
Results: In this work we propose a pipeline that couples a machine learning
solution for molecular profiling with a recent network comparison method. The
pipeline can identify changes occurring between specific sub-modules of
networks built in a case-control biomarker study, discriminating key groups of
genes whose interactions are modified by an underlying condition. The proposal
is independent from the classification algorithm used. Three applications on
genomewide data are presented regarding children susceptibility to air
pollution and two neurodegenerative diseases: Parkinson's and Alzheimer's.
Availability: Details about the software used for the experiments discussed
in this paper are provided in the Appendix
How Gibbs distributions may naturally arise from synaptic adaptation mechanisms. A model-based argumentation
This paper addresses two questions in the context of neuronal networks
dynamics, using methods from dynamical systems theory and statistical physics:
(i) How to characterize the statistical properties of sequences of action
potentials ("spike trains") produced by neuronal networks ? and; (ii) what are
the effects of synaptic plasticity on these statistics ? We introduce a
framework in which spike trains are associated to a coding of membrane
potential trajectories, and actually, constitute a symbolic coding in important
explicit examples (the so-called gIF models). On this basis, we use the
thermodynamic formalism from ergodic theory to show how Gibbs distributions are
natural probability measures to describe the statistics of spike trains, given
the empirical averages of prescribed quantities. As a second result, we show
that Gibbs distributions naturally arise when considering "slow" synaptic
plasticity rules where the characteristic time for synapse adaptation is quite
longer than the characteristic time for neurons dynamics.Comment: 39 pages, 3 figure
Stochastic pump effect and geometric phases in dissipative and stochastic systems
The success of Berry phases in quantum mechanics stimulated the study of
similar phenomena in other areas of physics, including the theory of living
cell locomotion and motion of patterns in nonlinear media. More recently,
geometric phases have been applied to systems operating in a strongly
stochastic environment, such as molecular motors. We discuss such geometric
effects in purely classical dissipative stochastic systems and their role in
the theory of the stochastic pump effect (SPE).Comment: Review. 35 pages. J. Phys. A: Math, Theor. (in press
Measures and Limits of Models of Fixation Selection
Models of fixation selection are a central tool in the quest to understand how the human mind selects relevant information. Using this tool in the evaluation of competing claims often requires comparing different models' relative performance in predicting eye movements. However, studies use a wide variety of performance measures with markedly different properties, which makes a comparison difficult. We make three main contributions to this line of research: First we argue for a set of desirable properties, review commonly used measures, and conclude that no single measure unites all desirable properties. However the area under the ROC curve (a classification measure) and the KL-divergence (a distance measure of probability distributions) combine many desirable properties and allow a meaningful comparison of critical model performance. We give an analytical proof of the linearity of the ROC measure with respect to averaging over subjects and demonstrate an appropriate correction of entropy-based measures like KL-divergence for small sample sizes in the context of eye-tracking data. Second, we provide a lower bound and an upper bound of these measures, based on image-independent properties of fixation data and between subject consistency respectively. Based on these bounds it is possible to give a reference frame to judge the predictive power of a model of fixation selection . We provide open-source python code to compute the reference frame. Third, we show that the upper, between subject consistency bound holds only for models that predict averages of subject populations. Departing from this we show that incorporating subject-specific viewing behavior can generate predictions which surpass that upper bound. Taken together, these findings lay out the required information that allow a well-founded judgment of the quality of any model of fixation selection and should therefore be reported when a new model is introduced
- …