18 research outputs found
Sharp estimation in sup norm with random design
The aim of this paper is to recover the regression function with sup norm
loss. We construct an asymptotically sharp estimator which converges with the
spatially dependent rate r\_{n, \mu}(x) = P \big(\log n / (n \mu(x)) \big)^{s /
(2s + 1)}, where is the design density, the regression smoothness,
the sample size and is a constant expressed in terms of a solution to a
problem of optimal recovery as in Donoho (1994). We prove this result under the
assumption that is positive and continuous. This estimator combines
kernel and local polynomial methods, where the kernel is given by optimal
recovery, which allows to prove the result up to the constants for any .
Moreover, the estimator does not depend on . We prove that is optimal in a sense which is stronger than the classical minimax
lower bound. Then, an inhomogeneous confidence band is proposed. This band has
a non constant length which depends on the local amount of data
Convergence rates for pointwise curve estimation with a degenerate design
The nonparametric regression with a random design model is considered. We
want to recover the regression function at a point x where the design density
is vanishing or exploding. Depending on assumptions on the regression function
local regularity and on the design local behaviour, we find several minimax
rates. These rates lie in a wide range, from slow l(n) rates where l(.) is
slowly varying (for instance (log n)^(-1)) to fast n^(-1/2) * l(n) rates. If
the continuity modulus of the regression function at x can be bounded from
above by a s-regularly varying function, and if the design density is
b-regularly varying, we prove that the minimax convergence rate at x is
n^(-s/(1+2s+b)) * l(n)
Link Prediction in Graphs with Autoregressive Features
In the paper, we consider the problem of link prediction in time-evolving
graphs. We assume that certain graph features, such as the node degree, follow
a vector autoregressive (VAR) model and we propose to use this information to
improve the accuracy of prediction. Our strategy involves a joint optimization
procedure over the space of adjacency matrices and VAR matrices which takes
into account both sparsity and low rank properties of the matrices. Oracle
inequalities are derived and illustrate the trade-offs in the choice of
smoothing parameters when modeling the joint effect of sparsity and low rank
property. The estimate is computed efficiently using proximal methods through a
generalized forward-backward agorithm.Comment: NIPS 201
Estimation for the Prediction of Point Processes with Many Covariates
Estimation of the intensity of a point process is considered within a
nonparametric framework. The intensity measure is unknown and depends on
covariates, possibly many more than the observed number of jumps. Only a single
trajectory of the counting process is observed. Interest lies in estimating the
intensity conditional on the covariates. The impact of the covariates is
modelled by an additive model where each component can be written as a linear
combination of possibly unknown functions. The focus is on prediction as
opposed to variable screening. Conditions are imposed on the coefficients of
this linear combination in order to control the estimation error. The rates of
convergence are optimal when the number of active covariates is large. As an
application, the intensity of the buy and sell trades of the New Zealand dollar
futures is estimated and a test for forecast evaluation is presented. A
simulation is included to provide some finite sample intuition on the model and
asymptotic properties
Régression non-paramétrique et information spatialement inhomogène
We study the nonparametric estimation of a signal based on inhomogeneous noisy data (the amount of data varies on the estimation domain). We consider the model of nonparametric regression with random design. Our aim is to understand the consequences of inhomogeneous data on the estimation problem in the minimax setup. Our approach is twofold: local and global. In the local setup, we want to recover the regression at a point with little, or much data. By translating this property into several assumptions on the design density, we obtain a large range of new minimax rates, containing very slow and very fast rates. Then, we construct a smoothness adaptive procedure, and we show that it converges with aminimax rate penalised by a minimal cost. In the global setup, we wantto recover the regression with sup norm loss. We propose estimatorsconverging with rates which are sensitive to the inhomogeneousbehaviour of the information in the model. We prove the spatialoptimality of these rates, which consists in an enforcement of theclassical minimax lower bound for sup norm loss. In particular, weconstruct an asymptotically sharp estimator over Hölder balls withany smoothness, and a confidence band with a width which adapts to thelocal amount of data.Nous étudions l'estimation non-paramétrique d'un signal à partir dedonnées bruitées spatialement inhomogènes (données dont la quantitévarie sur le domaine d'estimation). Le prototype d'étude est le modèlede régression avec design aléatoire. Notre objectif est de comprendreles conséquences du caractère inhomogène des données sur le problèmed'estimation dans le cadre d'étude minimax. Nous adoptons deux pointsde vue : local et global. Du point de vue local, nous nous intéressonsà l'estimation de la régression en un point avec peu ou beaucoup dedonnées. En traduisant cette propriété par différentes hypothèses surle comportement local de la densité du design, nous obtenons toute unegamme de nouvelles vitesses minimax ponctuelles, comprenant desvitesses très lentes et des vitesses très rapides. Puis, nousconstruisons une procédure adaptative en la régularité de larégression, et nous montrons qu'elle converge avec la vitesse minimaxà laquelle s'ajoute un coût minimal pour l'adaptation locale. Du pointde vue global, nous nous intéressons à l'estimation de la régressionen perte uniforme. Nous proposons des estimateurs qui convergent avecdes vitesses dépendantes de l'espace, lesquelles rendent compte ducaractère inhomogène de l'information dans le modèle. Nous montronsl'optimalité spatiale de ces vitesses, qui consiste en un renforcementde la borne inférieure minimax classique pour la perte uniforme. Nousconstruisons notamment un estimateur asymptotiquement exact sur uneboule de Hölder de régularité quelconque, ainsi qu'une bande deconfiance dont la largeur s'adapte à la quantité locale de données
High dimensional matrix estimation with unknown variance of the noise
International audienceWe propose a new pivotal method for estimating high-dimensional matrices. Assume that we observe a small set of entries or linear combinations of entries of an unknown matrix corrupted by noise. We propose a new method for estimating which does not rely on the knowledge or an estimation of the standard deviation of the noise . Our estimator achieves, up to a logarithmic factor, optimal rates of convergence under the Frobenius risk and, thus, has the same prediction performance as previously proposed estimators which rely on the knowledge of . Our method is based on the solution of a convex optimization problem which makes it computationally attractive
Inégalités d'oracle exactes pour la prédiction d'une matrice en grande dimension
International audienceWe consider the problem of prediction of a high dimensional matrix of size with noise, meaning that is much larger than the sample size . We focus on the trace norm minimization algorithm, but also on other penalizations. It is now well-known that such algorithms can be used for matrix completion, as well as other problems, such as multi-task learning, see \cite{candes-plan2,candes-recht08,candes-plan1,candes-tao1, rohde-tsyb09, MR2417263}. In this work, we propose sharp oracle inequalities in a statistical learning setup