4,200 research outputs found
Conditional independence testing based on a nearest-neighbor estimator of conditional mutual information
Conditional independence testing is a fundamental problem underlying causal
discovery and a particularly challenging task in the presence of nonlinear and
high-dimensional dependencies. Here a fully non-parametric test for continuous
data based on conditional mutual information combined with a local permutation
scheme is presented. Through a nearest neighbor approach, the test efficiently
adapts also to non-smooth distributions due to strongly nonlinear dependencies.
Numerical experiments demonstrate that the test reliably simulates the null
distribution even for small sample sizes and with high-dimensional conditioning
sets. The test is better calibrated than kernel-based tests utilizing an
analytical approximation of the null distribution, especially for non-smooth
densities, and reaches the same or higher power levels. Combining the local
permutation scheme with the kernel tests leads to better calibration, but
suffers in power. For smaller sample sizes and lower dimensions, the test is
faster than random fourier feature-based kernel tests if the permutation scheme
is (embarrassingly) parallelized, but the runtime increases more sharply with
sample size and dimensionality. Thus, more theoretical research to analytically
approximate the null distribution and speed up the estimation for larger sample
sizes is desirable.Comment: 17 pages, 12 figures, 1 tabl
Ensemble Kalman methods for high-dimensional hierarchical dynamic space-time models
We propose a new class of filtering and smoothing methods for inference in
high-dimensional, nonlinear, non-Gaussian, spatio-temporal state-space models.
The main idea is to combine the ensemble Kalman filter and smoother, developed
in the geophysics literature, with state-space algorithms from the statistics
literature. Our algorithms address a variety of estimation scenarios, including
on-line and off-line state and parameter estimation. We take a Bayesian
perspective, for which the goal is to generate samples from the joint posterior
distribution of states and parameters. The key benefit of our approach is the
use of ensemble Kalman methods for dimension reduction, which allows inference
for high-dimensional state vectors. We compare our methods to existing ones,
including ensemble Kalman filters, particle filters, and particle MCMC. Using a
real data example of cloud motion and data simulated under a number of
nonlinear and non-Gaussian scenarios, we show that our approaches outperform
these existing methods
Sparse Extended Information Filters: Insights into Sparsification
Recently, there have been a number of variant
Simultaneous Localization and Mapping (SLAM) algorithms that
have made substantial progress towards large-area scalability
by parameterizing the SLAM posterior within the information
(canonical/inverse covariance) form. Of these, probably the most
well-known and popular approach is the Sparse Extended Information
Filter (SEIF) by Thrun et al. While SEIFs have been
successfully implemented with a variety of challenging real-world
datasets and have led to new insights into scalable SLAM, open
research questions remain regarding the approximate sparsification
procedure and its effect on map error consistency.
In this paper, we examine the constant-time SEIF sparsification
procedure in depth and offer new insight into issues of consistency.
In particular, we show that exaggerated map inconsistency
occurs within the global reference frame where estimation is
performed, but that empirical testing shows that relative local
map relationships are preserved. We then present a slightly
modified version of their sparsification procedure, which is shown
to preserve sparsity while also generating both local and global
map estimates comparable to those obtained by the non-sparsified
SLAM filter. While this modified approximation is no longer
constant-time, it does serve as a theoretical benchmark against
which to compare SEIFs constant-time results. We demonstrate
our findings by benchmark comparison of the modified and
original SEIF sparsification rule using simulation in the linear
Gaussian SLAM case and real-world experiments for a nonlinear dataset.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/86045/1/reustice-31.pd
- …