2,671 research outputs found
Bayesian inference through encompassing priors and importance sampling for a class of marginal models for categorical data
We develop a Bayesian approach for selecting the model which is the most
supported by the data within a class of marginal models for categorical
variables formulated through equality and/or inequality constraints on
generalised logits (local, global, continuation or reverse continuation),
generalised log-odds ratios and similar higher-order interactions. For each
constrained model, the prior distribution of the model parameters is formulated
following the encompassing prior approach. Then, model selection is performed
by using Bayes factors which are estimated by an importance sampling method.
The approach is illustrated through three applications involving some datasets,
which also include explanatory variables. In connection with one of these
examples, a sensitivity analysis to the prior specification is also considered
Euler integration over definable functions
We extend the theory of Euler integration from the class of constructible
functions to that of "tame" real-valued functions (definable with respect to an
o-minimal structure). The corresponding integral operator has some unusual
defects (it is not a linear operator); however, it has a compelling
Morse-theoretic interpretation. In addition, we show that it is an appropriate
setting in which to do numerical analysis of Euler integrals, with applications
to incomplete and uncertain data in sensor networks.Comment: 6 page
Intervention analysis with state-space models to estimate discontinuities due to a survey redesign
An important quality aspect of official statistics produced by national
statistical institutes is comparability over time. To maintain uninterrupted
time series, surveys conducted by national statistical institutes are often
kept unchanged as long as possible. To improve the quality or efficiency of a
survey process, however, it remains inevitable to adjust methods or redesign
this process from time to time. Adjustments in the survey process generally
affect survey characteristics such as response bias and therefore have a
systematic effect on the parameter estimates of a sample survey. Therefore, it
is important that the effects of a survey redesign on the estimated series are
explained and quantified. In this paper a structural time series model is
applied to estimate discontinuities in series of the Dutch survey on social
participation and environmental consciousness due to a redesign of the
underlying survey process.Comment: Published in at http://dx.doi.org/10.1214/09-AOAS305 the Annals of
Applied Statistics (http://www.imstat.org/aoas/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Robust estimation of risks from small samples
Data-driven risk analysis involves the inference of probability distributions
from measured or simulated data. In the case of a highly reliable system, such
as the electricity grid, the amount of relevant data is often exceedingly
limited, but the impact of estimation errors may be very large. This paper
presents a robust nonparametric Bayesian method to infer possible underlying
distributions. The method obtains rigorous error bounds even for small samples
taken from ill-behaved distributions. The approach taken has a natural
interpretation in terms of the intervals between ordered observations, where
allocation of probability mass across intervals is well-specified, but the
location of that mass within each interval is unconstrained. This formulation
gives rise to a straightforward computational resampling method: Bayesian
Interval Sampling. In a comparison with common alternative approaches, it is
shown to satisfy strict error bounds even for ill-behaved distributions.Comment: 13 pages, 3 figures; supplementary information provided. A revised
version of this manuscript has been accepted for publication in Philosophical
Transactions of the Royal Society A: Mathematical, Physical and Engineering
Science
Estimating Discrete Markov Models From Various Incomplete Data Schemes
The parameters of a discrete stationary Markov model are transition
probabilities between states. Traditionally, data consist in sequences of
observed states for a given number of individuals over the whole observation
period. In such a case, the estimation of transition probabilities is
straightforwardly made by counting one-step moves from a given state to
another. In many real-life problems, however, the inference is much more
difficult as state sequences are not fully observed, namely the state of each
individual is known only for some given values of the time variable. A review
of the problem is given, focusing on Monte Carlo Markov Chain (MCMC) algorithms
to perform Bayesian inference and evaluate posterior distributions of the
transition probabilities in this missing-data framework. Leaning on the
dependence between the rows of the transition matrix, an adaptive MCMC
mechanism accelerating the classical Metropolis-Hastings algorithm is then
proposed and empirically studied.Comment: 26 pages - preprint accepted in 20th February 2012 for publication in
Computational Statistics and Data Analysis (please cite the journal's paper
- …