48 research outputs found
Bayesian Synthesis: Combining subjective analyses, with an application to ozone data
Bayesian model averaging enables one to combine the disparate predictions of
a number of models in a coherent fashion, leading to superior predictive
performance. The improvement in performance arises from averaging models that
make different predictions. In this work, we tap into perhaps the biggest
driver of different predictions---different analysts---in order to gain the
full benefits of model averaging. In a standard implementation of our method,
several data analysts work independently on portions of a data set, eliciting
separate models which are eventually updated and combined through a specific
weighting method. We call this modeling procedure Bayesian Synthesis. The
methodology helps to alleviate concerns about the sizable gap between the
foundational underpinnings of the Bayesian paradigm and the practice of
Bayesian statistics. In experimental work we show that human modeling has
predictive performance superior to that of many automatic modeling techniques,
including AIC, BIC, Smoothing Splines, CART, Bagged CART, Bayes CART, BMA and
LARS, and only slightly inferior to that of BART. We also show that Bayesian
Synthesis further improves predictive performance. Additionally, we examine the
predictive performance of a simple average across analysts, which we dub Convex
Synthesis, and find that it also produces an improvement.Comment: Published in at http://dx.doi.org/10.1214/10-AOAS444 the Annals of
Applied Statistics (http://www.imstat.org/aoas/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Rediscovering a little known fact about the t-test and the F-test: Algebraic, Geometric, Distributional and Graphical Considerations
We discuss the role that the null hypothesis should play in the construction
of a test statistic used to make a decision about that hypothesis. To construct
the test statistic for a point null hypothesis about a binomial proportion, a
common recommendation is to act as if the null hypothesis is true. We argue
that, on the surface, the one-sample t-test of a point null hypothesis about a
Gaussian population mean does not appear to follow the recommendation. We show
how simple algebraic manipulations of the usual t-statistic lead to an
equivalent test procedure consistent with the recommendation. We provide
geometric intuition regarding this equivalence and we consider extensions to
testing nested hypotheses in Gaussian linear models. We discuss an application
to graphical residual diagnostics where the form of the test statistic makes a
practical difference. By examining the formulation of the test statistic from
multiple perspectives in this familiar example, we provide simple, concrete
illustrations of some important issues that can guide the formulation of
effective solutions to more complex statistical problems.Comment: 22 pages, 5 figure
Bayesian Restricted Likelihood Methods: Conditioning on Insufficient Statistics in Bayesian Regression
Bayesian methods have proven themselves to be successful across a wide range
of scientific problems and have many well-documented advantages over competing
methods. However, these methods run into difficulties for two major and
prevalent classes of problems: handling data sets with outliers and dealing
with model misspecification. We outline the drawbacks of previous solutions to
both of these problems and propose a new method as an alternative. When working
with the new method, the data is summarized through a set of insufficient
statistics, targeting inferential quantities of interest, and the prior
distribution is updated with the summary statistics rather than the complete
data. By careful choice of conditioning statistics, we retain the main benefits
of Bayesian methods while reducing the sensitivity of the analysis to features
of the data not captured by the conditioning statistics. For reducing
sensitivity to outliers, classical robust estimators (e.g., M-estimators) are
natural choices for conditioning statistics. A major contribution of this work
is the development of a data augmented Markov chain Monte Carlo (MCMC)
algorithm for the linear model and a large class of summary statistics. We
demonstrate the method on simulated and real data sets containing outliers and
subject to model misspecification. Success is manifested in better predictive
performance for data points of interest as compared to competing methods
A Process of Dependent Quantile Pyramids
Despite the practicality of quantile regression (QR), simultaneous estimation
of multiple QR curves continues to be challenging. We address this problem by
proposing a Bayesian nonparametric framework that generalizes the quantile
pyramid by replacing each scalar variate in the quantile pyramid with a
stochastic process on a covariate space. We propose a novel approach to show
the existence of a quantile pyramid for all quantiles. The process of dependent
quantile pyramids allows for non-linear QR and automatically ensures
non-crossing of QR curves on the covariate space. Simulation studies document
the performance and robustness of our approach. An application to cyclone
intensity data is presented.Comment: 36 pages, 4 figures, 2 table
Efficient Model Selection in Linear and Non-Linear Quantile Regression by Cross-Validation
Check loss function is used to define quantile regression. In the prospect of cross validation, it is also employed as a validation function when underlying truth is unknown. However, our empirical study indicates that the validation with check loss often leads to choosing an over estimated fits. In this work, we suggest a modified or L2-adjusted check loss which rounds the sharp corner in the middle of check loss. It has a large effect of guarding against over fitted model in some extent. Through various simulation settings of linear and non-linear regressions, the improvement of check loss by L2 adjustment is empirically examined. This adjustment is devised to shrink to zero as sample size grows
On the two-dataset problem
This paper considers the two-dataset problem, where data are collected from
two potentially different populations sharing common aspects. This problem
arises when data are collected by two different types of researchers or from
two different sources. We may reach invalid conclusions without using knowledge
about the data collection process. To address this problem, this paper develops
statistical models focusing on the difference in measurement and proposes two
prediction errors that help to evaluate the underlying data collection process.
As a consequence, it is possible to discuss the heterogeneity/similarity of
data in terms of prediction. Two real datasets are selected to illustrate our
method