8,845 research outputs found
Brain covariance selection: better individual functional connectivity models using population prior
Spontaneous brain activity, as observed in functional neuroimaging, has been
shown to display reproducible structure that expresses brain architecture and
carries markers of brain pathologies. An important view of modern neuroscience
is that such large-scale structure of coherent activity reflects modularity
properties of brain connectivity graphs. However, to date, there has been no
demonstration that the limited and noisy data available in spontaneous activity
observations could be used to learn full-brain probabilistic models that
generalize to new data. Learning such models entails two main challenges: i)
modeling full brain connectivity is a difficult estimation problem that faces
the curse of dimensionality and ii) variability between subjects, coupled with
the variability of functional signals between experimental runs, makes the use
of multiple datasets challenging. We describe subject-level brain functional
connectivity structure as a multivariate Gaussian process and introduce a new
strategy to estimate it from group data, by imposing a common structure on the
graphical model in the population. We show that individual models learned from
functional Magnetic Resonance Imaging (fMRI) data using this population prior
generalize better to unseen data than models based on alternative
regularization schemes. To our knowledge, this is the first report of a
cross-validated model of spontaneous brain activity. Finally, we use the
estimated graphical model to explore the large-scale characteristics of
functional architecture and show for the first time that known cognitive
networks appear as the integrated communities of functional connectivity graph.Comment: in Advances in Neural Information Processing Systems, Vancouver :
Canada (2010
Information flow between resting state networks
The resting brain dynamics self-organizes into a finite number of correlated
patterns known as resting state networks (RSNs). It is well known that
techniques like independent component analysis can separate the brain activity
at rest to provide such RSNs, but the specific pattern of interaction between
RSNs is not yet fully understood. To this aim, we propose here a novel method
to compute the information flow (IF) between different RSNs from resting state
magnetic resonance imaging. After haemodynamic response function blind
deconvolution of all voxel signals, and under the hypothesis that RSNs define
regions of interest, our method first uses principal component analysis to
reduce dimensionality in each RSN to next compute IF (estimated here in terms
of Transfer Entropy) between the different RSNs by systematically increasing k
(the number of principal components used in the calculation). When k = 1, this
method is equivalent to computing IF using the average of all voxel activities
in each RSN. For k greater than one our method calculates the k-multivariate IF
between the different RSNs. We find that the average IF among RSNs is
dimension-dependent, increasing from k =1 (i.e., the average voxels activity)
up to a maximum occurring at k =5 to finally decay to zero for k greater than
10. This suggests that a small number of components (close to 5) is sufficient
to describe the IF pattern between RSNs. Our method - addressing differences in
IF between RSNs for any generic data - can be used for group comparison in
health or disease. To illustrate this, we have calculated the interRSNs IF in a
dataset of Alzheimer's Disease (AD) to find that the most significant
differences between AD and controls occurred for k =2, in addition to AD
showing increased IF w.r.t. controls.Comment: 47 pages, 5 figures, 4 tables, 3 supplementary figures. Accepted for
publication in Brain Connectivity in its current for
Local-Aggregate Modeling for Big-Data via Distributed Optimization: Applications to Neuroimaging
Technological advances have led to a proliferation of structured big data
that have matrix-valued covariates. We are specifically motivated to build
predictive models for multi-subject neuroimaging data based on each subject's
brain imaging scans. This is an ultra-high-dimensional problem that consists of
a matrix of covariates (brain locations by time points) for each subject; few
methods currently exist to fit supervised models directly to this tensor data.
We propose a novel modeling and algorithmic strategy to apply generalized
linear models (GLMs) to this massive tensor data in which one set of variables
is associated with locations. Our method begins by fitting GLMs to each
location separately, and then builds an ensemble by blending information across
locations through regularization with what we term an aggregating penalty. Our
so called, Local-Aggregate Model, can be fit in a completely distributed manner
over the locations using an Alternating Direction Method of Multipliers (ADMM)
strategy, and thus greatly reduces the computational burden. Furthermore, we
propose to select the appropriate model through a novel sequence of faster
algorithmic solutions that is similar to regularization paths. We will
demonstrate both the computational and predictive modeling advantages of our
methods via simulations and an EEG classification problem.Comment: 41 pages, 5 figures and 3 table
- …