53,953 research outputs found
Learning Sets with Separating Kernels
We consider the problem of learning a set from random samples. We show how
relevant geometric and topological properties of a set can be studied
analytically using concepts from the theory of reproducing kernel Hilbert
spaces. A new kind of reproducing kernel, that we call separating kernel, plays
a crucial role in our study and is analyzed in detail. We prove a new analytic
characterization of the support of a distribution, that naturally leads to a
family of provably consistent regularized learning algorithms and we discuss
the stability of these methods with respect to random sampling. Numerical
experiments show that the approach is competitive, and often better, than other
state of the art techniques.Comment: final versio
A note on conditional versus joint unconditional weak convergence in bootstrap consistency results
The consistency of a bootstrap or resampling scheme is classically validated
by weak convergence of conditional laws. However, when working with stochastic
processes in the space of bounded functions and their weak convergence in the
Hoffmann-J{\o}rgensen sense, an obstacle occurs: due to possible
non-measurability, neither laws nor conditional laws are well-defined. Starting
from an equivalent formulation of weak convergence based on the bounded
Lipschitz metric, a classical circumvent is to formulate bootstrap consistency
in terms of the latter distance between what might be called a
\emph{conditional law} of the (non-measurable) bootstrap process and the law of
the limiting process. The main contribution of this note is to provide an
equivalent formulation of bootstrap consistency in the space of bounded
functions which is more intuitive and easy to work with. Essentially, the
equivalent formulation consists of (unconditional) weak convergence of the
original process jointly with two bootstrap replicates. As a by-product, we
provide two equivalent formulations of bootstrap consistency for statistics
taking values in separable metric spaces: the first in terms of (unconditional)
weak convergence of the statistic jointly with its bootstrap replicates, the
second in terms of convergence in probability of the empirical distribution
function of the bootstrap replicates. Finally, the asymptotic validity of
bootstrap-based confidence intervals and tests is briefly revisited, with
particular emphasis on the, in practice unavoidable, Monte Carlo approximation
of conditional quantiles.Comment: 21 pages, 1 Figur
On choosing and bounding probability metrics
When studying convergence of measures, an important issue is the choice of
probability metric. In this review, we provide a summary and some new results
concerning bounds among ten important probability metrics/distances that are
used by statisticians and probabilists. We focus on these metrics because they
are either well-known, commonly used, or admit practical bounding techniques.
We summarize these relationships in a handy reference diagram, and also give
examples to show how rates of convergence can depend on the metric chosen.Comment: To appear, International Statistical Review. Related work at
http://www.math.hmc.edu/~su/papers.htm
Differentiable structures on metric measure spaces: A Primer
This is an exposition of the theory of differentiable structures on metric
measures spaces, in the sense of Cheeger and Keith.Comment: 23 page
- …