24,075 research outputs found
Evaluation of second-level inference in fMRI analysis
We investigate the impact of decisions in the second-level (i.e., over subjects) inferential process in functional magnetic resonance imaging on (1) the balance between false positives and false negatives and on (2) the data-analytical stability, both proxies for the reproducibility of results. Second-level analysis based on a mass univariate approach typically consists of 3 phases. First, one proceeds via a general linear model for a test image that consists of pooled information from different subjects. We evaluate models that take into account first-level (within-subjects) variability and models that do not take into account this variability. Second, one proceeds via inference based on parametrical assumptions or via permutation-based inference. Third, we evaluate 3 commonly used procedures to address the multiple testing problem: familywise error rate correction, False Discovery Rate (FDR) correction, and a two-step procedure with minimal cluster size. Based on a simulation study and real data we find that the two-step procedure with minimal cluster size results in most stable results, followed by the familywise error rate correction. The FDR results in most variable results, for both permutation-based inference and parametrical inference. Modeling the subject-specific variability yields a better balance between false positives and false negatives when using parametric inference
Permutation Inference for Canonical Correlation Analysis
Canonical correlation analysis (CCA) has become a key tool for population
neuroimaging, allowing investigation of associations between many imaging and
non-imaging measurements. As other variables are often a source of variability
not of direct interest, previous work has used CCA on residuals from a model
that removes these effects, then proceeded directly to permutation inference.
We show that such a simple permutation test leads to inflated error rates. The
reason is that residualisation introduces dependencies among the observations
that violate the exchangeability assumption. Even in the absence of nuisance
variables, however, a simple permutation test for CCA also leads to excess
error rates for all canonical correlations other than the first. The reason is
that a simple permutation scheme does not ignore the variability already
explained by previous canonical variables. Here we propose solutions for both
problems: in the case of nuisance variables, we show that transforming the
residuals to a lower dimensional basis where exchangeability holds results in a
valid permutation test; for more general cases, with or without nuisance
variables, we propose estimating the canonical correlations in a stepwise
manner, removing at each iteration the variance already explained, while
dealing with different number of variables in both sides. We also discuss how
to address the multiplicity of tests, proposing an admissible test that is not
conservative, and provide a complete algorithm for permutation inference for
CCA.Comment: 49 pages, 2 figures, 10 tables, 3 algorithms, 119 reference
The Augmented Synthetic Control Method
The synthetic control method (SCM) is a popular approach for estimating the
impact of a treatment on a single unit in panel data settings. The "synthetic
control" is a weighted average of control units that balances the treated
unit's pre-treatment outcomes as closely as possible. A critical feature of the
original proposal is to use SCM only when the fit on pre-treatment outcomes is
excellent. We propose Augmented SCM as an extension of SCM to settings where
such pre-treatment fit is infeasible. Analogous to bias correction for inexact
matching, Augmented SCM uses an outcome model to estimate the bias due to
imperfect pre-treatment fit and then de-biases the original SCM estimate. Our
main proposal, which uses ridge regression as the outcome model, directly
controls pre-treatment fit while minimizing extrapolation from the convex hull.
This estimator can also be expressed as a solution to a modified synthetic
controls problem that allows negative weights on some donor units. We bound the
estimation error of this approach under different data generating processes,
including a linear factor model, and show how regularization helps to avoid
over-fitting to noise. We demonstrate gains from Augmented SCM with extensive
simulation studies and apply this framework to estimate the impact of the 2012
Kansas tax cuts on economic growth. We implement the proposed method in the new
augsynth R package
Faster Family-wise Error Control for Neuroimaging with a Parametric Bootstrap
In neuroimaging, hundreds to hundreds of thousands of tests are performed
across a set of brain regions or all locations in an image. Recent studies have
shown that the most common family-wise error (FWE) controlling procedures in
imaging, which rely on classical mathematical inequalities or Gaussian random
field theory, yield FWE rates that are far from the nominal level. Depending on
the approach used, the FWER can be exceedingly small or grossly inflated. Given
the widespread use of neuroimaging as a tool for understanding neurological and
psychiatric disorders, it is imperative that reliable multiple testing
procedures are available. To our knowledge, only permutation joint testing
procedures have been shown to reliably control the FWER at the nominal level.
However, these procedures are computationally intensive due to the increasingly
available large sample sizes and dimensionality of the images, and analyses can
take days to complete. Here, we develop a parametric bootstrap joint testing
procedure. The parametric bootstrap procedure works directly with the test
statistics, which leads to much faster estimation of adjusted \emph{p}-values
than resampling-based procedures while reliably controlling the FWER in sample
sizes available in many neuroimaging studies. We demonstrate that the procedure
controls the FWER in finite samples using simulations, and present region- and
voxel-wise analyses to test for sex differences in developmental trajectories
of cerebral blood flow
Disentangling causal webs in the brain using functional Magnetic Resonance Imaging: A review of current approaches
In the past two decades, functional Magnetic Resonance Imaging has been used
to relate neuronal network activity to cognitive processing and behaviour.
Recently this approach has been augmented by algorithms that allow us to infer
causal links between component populations of neuronal networks. Multiple
inference procedures have been proposed to approach this research question but
so far, each method has limitations when it comes to establishing whole-brain
connectivity patterns. In this work, we discuss eight ways to infer causality
in fMRI research: Bayesian Nets, Dynamical Causal Modelling, Granger Causality,
Likelihood Ratios, LiNGAM, Patel's Tau, Structural Equation Modelling, and
Transfer Entropy. We finish with formulating some recommendations for the
future directions in this area
Can parametric statistical methods be trusted for fMRI based group studies?
The most widely used task fMRI analyses use parametric methods that depend on
a variety of assumptions. While individual aspects of these fMRI models have
been evaluated, they have not been evaluated in a comprehensive manner with
empirical data. In this work, a total of 2 million random task fMRI group
analyses have been performed using resting state fMRI data, to compute
empirical familywise error rates for the software packages SPM, FSL and AFNI,
as well as a standard non-parametric permutation method. While there is some
variation, for a nominal familywise error rate of 5% the parametric statistical
methods are shown to be conservative for voxel-wise inference and invalid for
cluster-wise inference; in particular, cluster size inference with a cluster
defining threshold of p = 0.01 generates familywise error rates up to 60%. We
conduct a number of follow up analyses and investigations that suggest the
cause of the invalid cluster inferences is spatial auto correlation functions
that do not follow the assumed Gaussian shape. By comparison, the
non-parametric permutation test, which is based on a small number of
assumptions, is found to produce valid results for voxel as well as cluster
wise inference. Using real task data, we compare the results between one
parametric method and the permutation test, and find stark differences in the
conclusions drawn between the two using cluster inference. These findings speak
to the need of validating the statistical methods being used in the
neuroimaging field
- …