Adaptive nonparametric estimation of a multivariate regression function

Abstract

We consider the kernel estimation of a multivariate regression function at a point. Theoretical choices of the bandwidth are possible for attaining minimum mean squared error or for local scaling, in the sense of asymptotic distribution. However, these choices are not available in practice. We follow the approach of Krieger and Pickands (Ann. Statist.9 (1981) 1066-1078) and Abramson (J. Multivariate Anal.12 (1982), 562-567) in constructing adaptive estimates after demonstrating the weak convergence of some error process. As consequences, efficient data-driven consistent estimation is feasible, and data-driven local scaling is also feasible. In the latter instance, nearest-neighbor-type estimates and variance-stabilizing estimates are obtained as special cases.multivariate kernel regression estimation bias variance asymptotic normality mean square error tightness weak convergence in C[a,b] Gaussian process adaptation

    Similar works

    Full text

    thumbnail-image

    Available Versions

    Last time updated on 06/07/2012