794 research outputs found
High-Dimensional Bayesian Geostatistics
With the growing capabilities of Geographic Information Systems (GIS) and
user-friendly software, statisticians today routinely encounter geographically
referenced data containing observations from a large number of spatial
locations and time points. Over the last decade, hierarchical spatiotemporal
process models have become widely deployed statistical tools for researchers to
better understand the complex nature of spatial and temporal variability.
However, fitting hierarchical spatiotemporal models often involves expensive
matrix computations with complexity increasing in cubic order for the number of
spatial locations and temporal points. This renders such models unfeasible for
large data sets. This article offers a focused review of two methods for
constructing well-defined highly scalable spatiotemporal stochastic processes.
Both these processes can be used as "priors" for spatiotemporal random fields.
The first approach constructs a low-rank process operating on a
lower-dimensional subspace. The second approach constructs a Nearest-Neighbor
Gaussian Process (NNGP) that ensures sparse precision matrices for its finite
realizations. Both processes can be exploited as a scalable prior embedded
within a rich hierarchical modeling framework to deliver full Bayesian
inference. These approaches can be described as model-based solutions for big
spatiotemporal datasets. The models ensure that the algorithmic complexity has
floating point operations (flops), where the number of spatial
locations (per iteration). We compare these methods and provide some insight
into their methodological underpinnings
Bayesian Anal
With the growing capabilities of Geographic Information Systems (GIS) and user-friendly software, statisticians today routinely encounter geographically referenced data containing observations from a large number of spatial locations and time points. Over the last decade, hierarchical spatiotemporal process models have become widely deployed statistical tools for researchers to better understand the complex nature of spatial and temporal variability. However, fitting hierarchical spatiotemporal models often involves expensive matrix computations with complexity increasing in cubic order for the number of spatial locations and temporal points. This renders such models unfeasible for large data sets. This article offers a focused review of two methods for constructing well-defined highly scalable spatiotemporal stochastic processes. Both these processes can be used as "priors" for spatiotemporal random fields. The first approach constructs a low-rank process operating on a lower-dimensional subspace. The second approach constructs a Nearest-Neighbor Gaussian Process (NNGP) that ensures sparse precision matrices for its finite realizations. Both processes can be exploited as a scalable prior embedded within a rich hierarchical modeling framework to deliver full Bayesian inference. These approaches can be described as model-based solutions for big spatiotemporal datasets. The models ensure that the algorithmic complexity has ~ | floating point operations (flops), where | the number of spatial locations (per iteration). We compare these methods and provide some insight into their methodological underpinnings.R01 ES027027/ES/NIEHS NIH HHS/United StatesR01 OH010093/OH/NIOSH CDC HHS/United StatesRC1 GM092400/GM/NIGMS NIH HHS/United States2018-01-30T00:00:00Z29391920PMC5790125vault:2616
Graph-constrained Analysis for Multivariate Functional Data
Functional Gaussian graphical models (GGM) used for analyzing multivariate
functional data customarily estimate an unknown graphical model representing
the conditional relationships between the functional variables. However, in
many applications of multivariate functional data, the graph is known and
existing functional GGM methods cannot preserve a given graphical constraint.
In this manuscript, we demonstrate how to conduct multivariate functional
analysis that exactly conforms to a given inter-variable graph. We first show
the equivalence between partially separable functional GGM and graphical
Gaussian processes (GP), proposed originally for constructing optimal
covariance functions for multivariate spatial data that retain the conditional
independence relations in a given graphical model. The theoretical connection
help design a new algorithm that leverages Dempster's covariance selection to
calculate the maximum likelihood estimate of the covariance function for
multivariate functional data under graphical constraints. We also show that the
finite term truncation of functional GGM basis expansion used in practice is
equivalent to a low-rank graphical GP, which is known to oversmooth marginal
distributions. To remedy this, we extend our algorithm to better preserve
marginal distributions while still respecting the graph and retaining
computational scalability. The insights obtained from the new results presented
in this manuscript will help practitioners better understand the relationship
between these graphical models and in deciding on the appropriate method for
their specific multivariate data analysis task. The benefits of the proposed
algorithms are illustrated using empirical experiments and an application to
functional modeling of neuroimaging data using the connectivity graph among
regions of the brain.Comment: 23 pages, 6 figure
Massively parallel implicit equal-weights particle filter for ocean drift trajectory forecasting
Forecasting of ocean drift trajectories are important for many applications, including search and rescue operations, oil spill cleanup and iceberg risk mitigation. In an operational setting, forecasts of drift trajectories are produced based on computationally demanding forecasts of three-dimensional ocean currents. Herein, we investigate a complementary approach for shorter time scales by using the recently proposed two-stage implicit equal-weights particle filter applied to a simplified ocean model. To achieve this, we present a new algorithmic design for a data-assimilation system in which all components – including the model, model errors, and particle filter – take advantage of massively parallel compute architectures, such as graphical processing units. Faster computations can enable in-situ and ad-hoc model runs for emergency management, and larger ensembles for better uncertainty quantification. Using a challenging test case with near-realistic chaotic instabilities, we run data-assimilation experiments based on synthetic observations from drifting and moored buoys, and analyze the trajectory forecasts for the drifters. Our results show that even sparse drifter observations are sufficient to significantly improve short-term drift forecasts up to twelve hours. With equidistant moored buoys observing only 0.1% of the state space, the ensemble gives an accurate description of the true state after data assimilation followed by a high-quality probabilistic forecast
- …