391,954 research outputs found
Entropic Priors and Bayesian Model Selection
We demonstrate that the principle of maximum relative entropy (ME), used
judiciously, can ease the specification of priors in model selection problems.
The resulting effect is that models that make sharp predictions are
disfavoured, weakening the usual Bayesian "Occam's Razor". This is illustrated
with a simple example involving what Jaynes called a "sure thing" hypothesis.
Jaynes' resolution of the situation involved introducing a large number of
alternative "sure thing" hypotheses that were possible before we observed the
data. However, in more complex situations, it may not be possible to explicitly
enumerate large numbers of alternatives. The entropic priors formalism produces
the desired result without modifying the hypothesis space or requiring explicit
enumeration of alternatives; all that is required is a good model for the prior
predictive distribution for the data. This idea is illustrated with a simple
rigged-lottery example, and we outline how this idea may help to resolve a
recent debate amongst cosmologists: is dark energy a cosmological constant, or
has it evolved with time in some way? And how shall we decide, when the data
are in?Comment: Presented at MaxEnt 2009, the 29th International Workshop on Bayesian
Inference and Maximum Entropy Methods in Science and Engineering (July 5-10,
2009, Oxford, Mississippi, USA
Bayesian reordering model with feature selection
In phrase-based statistical machine translation systems, variation in grammatical structures between source and target languages can cause large movements of phrases. Modeling such movements is crucial in achieving translations of long sentences that appear natural in the target language. We explore generative learning approach to phrase reordering in Arabic to English. Formulating the reordering problem as a classification problem and using naive Bayes with feature selection, we achieve an improvement in the BLEU score over a lexicalized reordering model. The proposed model is compact, fast and scalable to a large corpus
The Dependence of Routine Bayesian Model Selection Methods on Irrelevant Alternatives
Bayesian methods - either based on Bayes Factors or BIC - are now widely used
for model selection. One property that might reasonably be demanded of any
model selection method is that if a model is preferred to a model
, when these two models are expressed as members of one model class
, this preference is preserved when they are embedded in a
different class . However, we illustrate in this paper that with
the usual implementation of these common Bayesian procedures this property does
not hold true even approximately. We therefore contend that to use these
methods it is first necessary for there to exist a "natural" embedding class.
We argue that in any context like the one illustrated in our running example of
Bayesian model selection of binary phylogenetic trees there is no such
embedding
Training samples in objective Bayesian model selection
Central to several objective approaches to Bayesian model selection is the
use of training samples (subsets of the data), so as to allow utilization of
improper objective priors. The most common prescription for choosing training
samples is to choose them to be as small as possible, subject to yielding
proper posteriors; these are called minimal training samples.
When data can vary widely in terms of either information content or impact on
the improper priors, use of minimal training samples can be inadequate.
Important examples include certain cases of discrete data, the presence of
censored observations, and certain situations involving linear models and
explanatory variables. Such situations require more sophisticated methods of
choosing training samples. A variety of such methods are developed in this
paper, and successfully applied in challenging situations
Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems
Approximate Bayesian computation methods can be used to evaluate posterior
distributions without having to calculate likelihoods. In this paper we discuss
and apply an approximate Bayesian computation (ABC) method based on sequential
Monte Carlo (SMC) to estimate parameters of dynamical models. We show that ABC
SMC gives information about the inferability of parameters and model
sensitivity to changes in parameters, and tends to perform better than other
ABC approaches. The algorithm is applied to several well known biological
systems, for which parameters and their credible intervals are inferred.
Moreover, we develop ABC SMC as a tool for model selection; given a range of
different mathematical descriptions, ABC SMC is able to choose the best model
using the standard Bayesian model selection apparatus.Comment: 26 pages, 9 figure
- …
