12,848 research outputs found
Generalized Methods of Trimmed Moments
High breakdown-point regression estimators protect against large errors and data contamination. We adapt and generalize the concept of trimming used by many of these robust estimators so that it can be employed in the context of the generalized method of moments. The proposed generalized method of trimmed moments (GMTM) offers a globally robust estimation approach (contrary to existing only locally robust estimators) applicable in econometric models identified and estimated using moment conditions. We derive the consistency and asymptotic distribution of GMTM in a general setting, propose a robust test of overidentifying conditions, and demonstrate the application of GMTM in the instrumental variable regression. We also compare the finite-sample performance of GMTM and existing estimators by means of Monte Carlo simulation.asymptotic normality;generalized method of moments;instrumental variables regression;robust estimation;trimming
Market depth and order size: an analysis of permanent price effects of DAX futures' trades
In this paper we empirically analyze the permanent price impact of trades by investigating the relation between unexpected net order flow and price changes. We use intraday data on German index futures. Our analysis based on a neural network model suggests that the assumption of a linear impact of orders on prices (which is often used in theoretical papers) is highly questionable. Therefore, empirical studies, comparing the depth of different markets, should be based on the whole price impact function instead of a simple ratio. To allow the market depth to depend on trade volume could open promising avenues for further theoretical research. This could lead to quite different trading strategies as in traditional models. --
Locally adaptive image denoising by a statistical multiresolution criterion
We demonstrate how one can choose the smoothing parameter in image denoising
by a statistical multiresolution criterion, both globally and locally. Using
inhomogeneous diffusion and total variation regularization as examples for
localized regularization schemes, we present an efficient method for locally
adaptive image denoising. As expected, the smoothing parameter serves as an
edge detector in this framework. Numerical examples illustrate the usefulness
of our approach. We also present an application in confocal microscopy
Smoothed L-estimation of Regression Function
The Nadaraya-Watson nonparametric estimator of regression is known to be highly sensitive to the presence of outliers in data.This sensitivity can be reduced, for example, by using local L-estimates of regression.Whereas the local L-estimation is traditionally done using an empirical conditional distribution function, we propose to use instead a smoothed conditional distribution function.The asymptotic distribution of the proposed estimator is derived under mild ¯-mixing conditions, and additionally, we show that the smoothed L-estimation approach provides computational as well as statistical ¯nite-sample improvements.Finally, the proposed method is applied to the modelling of implied volatilitynonparametric regression;L-estimation;smoothed cumulative distribution function
Recommended from our members
Semiparametric estimation for a class of time-inhomogenous diffusion processes
Copyright @ 2009 Institute of Statistical Science, Academia SinicaWe develop two likelihood-based approaches to semiparametrically estimate a class of time-inhomogeneous diffusion processes: log penalized splines (P-splines) and the local log-linear method. Positive volatility is naturally embedded and this positivity is not guaranteed in most existing diffusion models. We investigate different smoothing parameter selections. Separate bandwidths are used for drift and volatility estimation. In the log P-splines approach, different smoothness for different time varying coefficients is feasible by assigning different penalty parameters. We also provide theorems for both approaches and report statistical inference results. Finally, we present a case study using the weekly three-month Treasury bill data from 1954 to 2004. We find that the log P-splines approach seems to capture the volatility dip in mid-1960s the best. We also present an application to calculate a financial market risk measure called Value at Risk (VaR) using statistical estimates from log P-splines
On weighted local fitting and its relation to the Horvitz-Thompson estimator
Weighting is a largely used concept in many fields of statistics and has frequently caused controversies on its justification and profit. In this paper, we analyze a weighted version of the well-known local polynomial regression estimators, derive their asymptotic bias and variance, and find that the conflict between the asymptotically optimal weighting scheme and the practical requirements has a surprising counterpart in sampling theory, leading us back to the discussion on Basu's (1971) elephants
Marginal integration for nonparametric causal inference
We consider the problem of inferring the total causal effect of a single
variable intervention on a (response) variable of interest. We propose a
certain marginal integration regression technique for a very general class of
potentially nonlinear structural equation models (SEMs) with known structure,
or at least known superset of adjustment variables: we call the procedure
S-mint regression. We easily derive that it achieves the convergence rate as
for nonparametric regression: for example, single variable intervention effects
can be estimated with convergence rate assuming smoothness with
twice differentiable functions. Our result can also be seen as a major
robustness property with respect to model misspecification which goes much
beyond the notion of double robustness. Furthermore, when the structure of the
SEM is not known, we can estimate (the equivalence class of) the directed
acyclic graph corresponding to the SEM, and then proceed by using S-mint based
on these estimates. We empirically compare the S-mint regression method with
more classical approaches and argue that the former is indeed more robust, more
reliable and substantially simpler.Comment: 40 pages, 14 figure
A Semi-parametric Technique for the Quantitative Analysis of Dynamic Contrast-enhanced MR Images Based on Bayesian P-splines
Dynamic Contrast-enhanced Magnetic Resonance Imaging (DCE-MRI) is an
important tool for detecting subtle kinetic changes in cancerous tissue.
Quantitative analysis of DCE-MRI typically involves the convolution of an
arterial input function (AIF) with a nonlinear pharmacokinetic model of the
contrast agent concentration. Parameters of the kinetic model are biologically
meaningful, but the optimization of the non-linear model has significant
computational issues. In practice, convergence of the optimization algorithm is
not guaranteed and the accuracy of the model fitting may be compromised. To
overcome this problems, this paper proposes a semi-parametric penalized spline
smoothing approach, with which the AIF is convolved with a set of B-splines to
produce a design matrix using locally adaptive smoothing parameters based on
Bayesian penalized spline models (P-splines). It has been shown that kinetic
parameter estimation can be obtained from the resulting deconvolved response
function, which also includes the onset of contrast enhancement. Detailed
validation of the method, both with simulated and in vivo data, is provided
- …