2,708 research outputs found
Linear Mixed Models with Marginally Symmetric Nonparametric Random Effects
Linear mixed models (LMMs) are used as an important tool in the data analysis
of repeated measures and longitudinal studies. The most common form of LMMs
utilize a normal distribution to model the random effects. Such assumptions can
often lead to misspecification errors when the random effects are not normal.
One approach to remedy the misspecification errors is to utilize a point-mass
distribution to model the random effects; this is known as the nonparametric
maximum likelihood-fitted (NPML) model. The NPML model is flexible but requires
a large number of parameters to characterize the random-effects distribution.
It is often natural to assume that the random-effects distribution be at least
marginally symmetric. The marginally symmetric NPML (MSNPML) random-effects
model is introduced, which assumes a marginally symmetric point-mass
distribution for the random effects. Under the symmetry assumption, the MSNPML
model utilizes half the number of parameters to characterize the same number of
point masses as the NPML model; thus the model confers an advantage in economy
and parsimony. An EM-type algorithm is presented for the maximum likelihood
(ML) estimation of LMMs with MSNPML random effects; the algorithm is shown to
monotonically increase the log-likelihood and is proven to be convergent to a
stationary point of the log-likelihood function in the case of convergence.
Furthermore, it is shown that the ML estimator is consistent and asymptotically
normal under certain conditions, and the estimation of quantities such as the
random-effects covariance matrix and individual a posteriori expectations is
demonstrated
Estimation in Dirichlet random effects models
We develop a new Gibbs sampler for a linear mixed model with a Dirichlet
process random effect term, which is easily extended to a generalized linear
mixed model with a probit link function. Our Gibbs sampler exploits the
properties of the multinomial and Dirichlet distributions, and is shown to be
an improvement, in terms of operator norm and efficiency, over other commonly
used MCMC algorithms. We also investigate methods for the estimation of the
precision parameter of the Dirichlet process, finding that maximum likelihood
may not be desirable, but a posterior mode is a reasonable approach. Examples
are given to show how these models perform on real data. Our results complement
both the theoretical basis of the Dirichlet process nonparametric prior and the
computational work that has been done to date.Comment: Published in at http://dx.doi.org/10.1214/09-AOS731 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Bayesian Approximate Kernel Regression with Variable Selection
Nonlinear kernel regression models are often used in statistics and machine
learning because they are more accurate than linear models. Variable selection
for kernel regression models is a challenge partly because, unlike the linear
regression setting, there is no clear concept of an effect size for regression
coefficients. In this paper, we propose a novel framework that provides an
effect size analog of each explanatory variable for Bayesian kernel regression
models when the kernel is shift-invariant --- for example, the Gaussian kernel.
We use function analytic properties of shift-invariant reproducing kernel
Hilbert spaces (RKHS) to define a linear vector space that: (i) captures
nonlinear structure, and (ii) can be projected onto the original explanatory
variables. The projection onto the original explanatory variables serves as an
analog of effect sizes. The specific function analytic property we use is that
shift-invariant kernel functions can be approximated via random Fourier bases.
Based on the random Fourier expansion we propose a computationally efficient
class of Bayesian approximate kernel regression (BAKR) models for both
nonlinear regression and binary classification for which one can compute an
analog of effect sizes. We illustrate the utility of BAKR by examining two
important problems in statistical genetics: genomic selection (i.e. phenotypic
prediction) and association mapping (i.e. inference of significant variants or
loci). State-of-the-art methods for genomic selection and association mapping
are based on kernel regression and linear models, respectively. BAKR is the
first method that is competitive in both settings.Comment: 22 pages, 3 figures, 3 tables; theory added; new simulations
presented; references adde
Semiparametric inference in mixture models with predictive recursion marginal likelihood
Predictive recursion is an accurate and computationally efficient algorithm
for nonparametric estimation of mixing densities in mixture models. In
semiparametric mixture models, however, the algorithm fails to account for any
uncertainty in the additional unknown structural parameter. As an alternative
to existing profile likelihood methods, we treat predictive recursion as a
filter approximation to fitting a fully Bayes model, whereby an approximate
marginal likelihood of the structural parameter emerges and can be used for
inference. We call this the predictive recursion marginal likelihood.
Convergence properties of predictive recursion under model mis-specification
also lead to an attractive construction of this new procedure. We show
pointwise convergence of a normalized version of this marginal likelihood
function. Simulations compare the performance of this new marginal likelihood
approach that of existing profile likelihood methods as well as Dirichlet
process mixtures in density estimation. Mixed-effects models and an empirical
Bayes multiple testing application in time series analysis are also considered
Excludability: A laboratory study on forced ranking in team production
Exclusion has long been employed as a common disciplinary measure against defectors, both at work and in social life. In this paper, we study the effect of excludability - exclusion of the lowest contributor - on contributions in three different team production settings. We demonstrate theoretically and experimentally that excludability increases contributions. Excludability is particularly effective in production settings where the average or maximum effort determines team production. In these settings, we observe almost immediate convergence to full contribution. In settings where the minimum effort determines team production, excludability leads to a large increase in contributions only if the value of the excluded individual's contribution to the public good is redistributed among the included individuals
Consistency conditions for regulatory analysis of financial institutions: a comparison of frontier efficiency methods
We propose a set of consistency conditions that frontier efficiency measures should meet to be most useful for regulatory analysis or other purposes. The efficiency estimates should be consistent in their efficiency levels, rankings, and identification of best and worst firms, consistent over time and with competitive conditions in the market, and consistent with standard nonfrontier measures of performance. We provide evidence on these conditions by evaluating and comparing efficiency estimates on U.S. bank efficiency from variants of all four of the major approaches -- DEA, SFA, TFA, and DFA -- and find mixed results.Financial institutions ; Bank supervision
- …