6,163 research outputs found
A Bayesian Heteroscedastic GLM with Application to fMRI Data with Motion Spikes
We propose a voxel-wise general linear model with autoregressive noise and
heteroscedastic noise innovations (GLMH) for analyzing functional magnetic
resonance imaging (fMRI) data. The model is analyzed from a Bayesian
perspective and has the benefit of automatically down-weighting time points
close to motion spikes in a data-driven manner. We develop a highly efficient
Markov Chain Monte Carlo (MCMC) algorithm that allows for Bayesian variable
selection among the regressors to model both the mean (i.e., the design matrix)
and variance. This makes it possible to include a broad range of explanatory
variables in both the mean and variance (e.g., time trends, activation stimuli,
head motion parameters and their temporal derivatives), and to compute the
posterior probability of inclusion from the MCMC output. Variable selection is
also applied to the lags in the autoregressive noise process, making it
possible to infer the lag order from the data simultaneously with all other
model parameters. We use both simulated data and real fMRI data from OpenfMRI
to illustrate the importance of proper modeling of heteroscedasticity in fMRI
data analysis. Our results show that the GLMH tends to detect more brain
activity, compared to its homoscedastic counterpart, by allowing the variance
to change over time depending on the degree of head motion
Large-scale Heteroscedastic Regression via Gaussian Process
Heteroscedastic regression considering the varying noises among observations
has many applications in the fields like machine learning and statistics. Here
we focus on the heteroscedastic Gaussian process (HGP) regression which
integrates the latent function and the noise function together in a unified
non-parametric Bayesian framework. Though showing remarkable performance, HGP
suffers from the cubic time complexity, which strictly limits its application
to big data. To improve the scalability, we first develop a variational sparse
inference algorithm, named VSHGP, to handle large-scale datasets. Furthermore,
two variants are developed to improve the scalability and capability of VSHGP.
The first is stochastic VSHGP (SVSHGP) which derives a factorized evidence
lower bound, thus enhancing efficient stochastic variational inference. The
second is distributed VSHGP (DVSHGP) which (i) follows the Bayesian committee
machine formalism to distribute computations over multiple local VSHGP experts
with many inducing points; and (ii) adopts hybrid parameters for experts to
guard against over-fitting and capture local variety. The superiority of DVSHGP
and SVSHGP as compared to existing scalable heteroscedastic/homoscedastic GPs
is then extensively verified on various datasets.Comment: 14 pages, 15 figure
- …