10,445 research outputs found
Statistical Inference using the Morse-Smale Complex
The Morse-Smale complex of a function decomposes the sample space into
cells where is increasing or decreasing. When applied to nonparametric
density estimation and regression, it provides a way to represent, visualize,
and compare multivariate functions. In this paper, we present some statistical
results on estimating Morse-Smale complexes. This allows us to derive new
results for two existing methods: mode clustering and Morse-Smale regression.
We also develop two new methods based on the Morse-Smale complex: a
visualization technique for multivariate functions and a two-sample,
multivariate hypothesis test.Comment: 45 pages, 13 figures. Accepted to Electronic Journal of Statistic
Kernel functions based on triplet comparisons
Given only information in the form of similarity triplets "Object A is more
similar to object B than to object C" about a data set, we propose two ways of
defining a kernel function on the data set. While previous approaches construct
a low-dimensional Euclidean embedding of the data set that reflects the given
similarity triplets, we aim at defining kernel functions that correspond to
high-dimensional embeddings. These kernel functions can subsequently be used to
apply any kernel method to the data set
Bandwidth selection in kernel empirical risk minimization via the gradient
In this paper, we deal with the data-driven selection of multidimensional and
possibly anisotropic bandwidths in the general framework of kernel empirical
risk minimization. We propose a universal selection rule, which leads to
optimal adaptive results in a large variety of statistical models such as
nonparametric robust regression and statistical learning with errors in
variables. These results are stated in the context of smooth loss functions,
where the gradient of the risk appears as a good criterion to measure the
performance of our estimators. The selection rule consists of a comparison of
gradient empirical risks. It can be viewed as a nontrivial improvement of the
so-called Goldenshluger-Lepski method to nonlinear estimators. Furthermore, one
main advantage of our selection rule is the nondependency on the Hessian matrix
of the risk, usually involved in standard adaptive procedures.Comment: Published at http://dx.doi.org/10.1214/15-AOS1318 in the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
- …