3,229 research outputs found
A Note on the Estimation of the Hölder Constant
In this article, we develop a nonparametric estimator for the Hölder constant of a density function. We consider a simulation study to evaluate the performance of the proposal and construct smooth bootstrap confidence intervals. Also, we give a brief review over the impossibility to decide whether a density function is Hölder.Fil: Henry, Guillermo Sebastian. Universidad de Buenos Aires. Facultad de Ciencias Exactas y Naturales; Argentina. Consejo Nacional de Investigaciones CientÃficas y Técnicas; ArgentinaFil: Rodriguez, Daniela Andrea. Universidad de Buenos Aires. Facultad de Ciencias Exactas y Naturales; Argentina. Consejo Nacional de Investigaciones CientÃficas y Técnicas; ArgentinaFil: Sued, Raquel Mariela. Universidad de Buenos Aires. Facultad de Ciencias Exactas y Naturales; Argentina. Consejo Nacional de Investigaciones CientÃficas y Técnicas; Argentin
Scalable Hash-Based Estimation of Divergence Measures
We propose a scalable divergence estimation method based on hashing. Consider
two continuous random variables and whose densities have bounded
support. We consider a particular locality sensitive random hashing, and
consider the ratio of samples in each hash bin having non-zero numbers of Y
samples. We prove that the weighted average of these ratios over all of the
hash bins converges to f-divergences between the two samples sets. We show that
the proposed estimator is optimal in terms of both MSE rate and computational
complexity. We derive the MSE rates for two families of smooth functions; the
H\"{o}lder smoothness class and differentiable functions. In particular, it is
proved that if the density functions have bounded derivatives up to the order
, where is the dimension of samples, the optimal parametric MSE rate
of can be achieved. The computational complexity is shown to be
, which is optimal. To the best of our knowledge, this is the first
empirical divergence estimator that has optimal computational complexity and
achieves the optimal parametric MSE estimation rate.Comment: 11 pages, Proceedings of the 21st International Conference on
Artificial Intelligence and Statistics (AISTATS) 2018, Lanzarote, Spai
Direct Estimation of Information Divergence Using Nearest Neighbor Ratios
We propose a direct estimation method for R\'{e}nyi and f-divergence measures
based on a new graph theoretical interpretation. Suppose that we are given two
sample sets and , respectively with and samples, where
is a constant value. Considering the -nearest neighbor (-NN)
graph of in the joint data set , we show that the average powered
ratio of the number of points to the number of points among all -NN
points is proportional to R\'{e}nyi divergence of and densities. A
similar method can also be used to estimate f-divergence measures. We derive
bias and variance rates, and show that for the class of -H\"{o}lder
smooth functions, the estimator achieves the MSE rate of
. Furthermore, by using a weighted ensemble
estimation technique, for density functions with continuous and bounded
derivatives of up to the order , and some extra conditions at the support
set boundary, we derive an ensemble estimator that achieves the parametric MSE
rate of . Our estimators are more computationally tractable than other
competing estimators, which makes them appealing in many practical
applications.Comment: 2017 IEEE International Symposium on Information Theory (ISIT
Empirical Bayes conditional density estimation
The problem of nonparametric estimation of the conditional density of a
response, given a vector of explanatory variables, is classical and of
prominent importance in many prediction problems since the conditional density
provides a more comprehensive description of the association between the
response and the predictor than, for instance, does the regression function.
The problem has applications across different fields like economy, actuarial
sciences and medicine. We investigate empirical Bayes estimation of conditional
densities establishing that an automatic data-driven selection of the prior
hyper-parameters in infinite mixtures of Gaussian kernels, with
predictor-dependent mixing weights, can lead to estimators whose performance is
on par with that of frequentist estimators in being minimax-optimal (up to
logarithmic factors) rate adaptive over classes of locally H\"older smooth
conditional densities and in performing an adaptive dimension reduction if the
response is independent of (some of) the explanatory variables which,
containing no information about the response, are irrelevant to the purpose of
estimating its conditional density
Oscillating Gaussian Processes
In this article we introduce and study oscillating Gaussian processes defined
by , where
are free parameters and is either stationary or
self-similar Gaussian process. We study the basic properties of and we
consider estimation of the model parameters. In particular, we show that the
moment estimators converge in and are, when suitably normalised,
asymptotically normal
Minimal H\"older regularity implying finiteness of integral Menger curvature
We study two families of integral functionals indexed by a real number . One family is defined for 1-dimensional curves in and the other one
is defined for -dimensional manifolds in . These functionals are
described as integrals of appropriate integrands (strongly related to the
Menger curvature) raised to power . Given we prove that
regularity of the set (a curve or a manifold), with implies finiteness of both curvature functionals
( in the case of curves). We also show that is optimal by
constructing examples of functions with graphs of infinite
integral curvature
- …