435 research outputs found

    Model selection via Bayesian information capacity designs for generalised linear models

    Get PDF
    The first investigation is made of designs for screening experiments where the response variable is approximated by a generalised linear model. A Bayesian information capacity criterion is defined for the selection of designs that are robust to the form of the linear predictor. For binomial data and logistic regression, the effectiveness of these designs for screening is assessed through simulation studies using all-subsets regression and model selection via maximum penalised likelihood and a generalised information criterion. For Poisson data and log-linear regression, similar assessments are made using maximum likelihood and the Akaike information criterion for minimally-supported designs that are constructed analytically. The results show that effective screening, that is, high power with moderate type I error rate and false discovery rate, can be achieved through suitable choices for the number of design support points and experiment size. Logistic regression is shown to present a more challenging problem than log-linear regression. Some areas for future work are also indicated

    Robust designs for Poisson regression models

    Get PDF
    We consider the problem of how to construct robust designs for Poisson regression models. An analytical expression is derived for robust designs for first-order Poisson regression models where uncertainty exists in the prior parameter estimates. Given certain constraints in the methodology, it may be necessary to extend the robust designs for implementation in practical experiments. With these extensions, our methodology constructs designs which perform similarly, in terms of estimation, to current techniques, and offers the solution in a more timely manner. We further apply this analytic result to cases where uncertainty exists in the linear predictor. The application of this methodology to practical design problems such as screening experiments is explored. Given the minimal prior knowledge that is usually available when conducting such experiments, it is recommended to derive designs robust across a variety of systems. However, incorporating such uncertainty into the design process can be a computationally intense exercise. Hence, our analytic approach is explored as an alternative

    Modelling Survival Data to Account for Model Uncertainty: A Single Model or Model Averaging?

    Get PDF
    This study considered the problem of predicting survival, based on three alternative models: a single Weibull, a\ud mixture of Weibulls and a cure model. Instead of the common procedure of choosing a single ???best??? model, where\ud ???best??? is defined in terms of goodness of fit to the data, a Bayesian model averaging (BMA) approach was adopted to\ud account for model uncertainty. This was illustrated using a case study in which the aim was the description of\ud lymphoma cancer survival with covariates given by phenotypes and gene expression. The results of this study indicate\ud that if the sample size is sufficiently large, one of the three models emerge as having highest probability given the\ud data, as indicated by the goodness of fit measure; the Bayesian information criterion (BIC). However, when the sample\ud size was reduced, no single model was revealed as ???best???, suggesting that a BMA approach would be appropriate.\ud Although a BMA approach can compromise on goodness of fit to the data (when compared to the true model), it can\ud provide robust predictions and facilitate more detailed investigation of the relationships between gene expression\ud and patient survival

    An approach for finding fully Bayesian optimal designs using normal-based approximations to loss functions

    Get PDF
    The generation of decision-theoretic Bayesian optimal designs is complicated by the significant computational challenge of minimising an analytically intractable expected loss function over a, potentially, high-dimensional design space. A new general approach for approximately finding Bayesian optimal designs is proposed which uses computationally efficient normal-based approximations to posterior summaries to aid in approximating the expected loss. This new approach is demonstrated on illustrative, yet challenging, examples including hierarchical models for blocked experiments, and experimental aims of parameter estimation and model discrimination. Where possible, the results of the proposed methodology are compared, both in terms of performance and computing time, to results from using computationally more expensive, but potentially more accurate, Monte Carlo approximations. Moreover, the methodology is also applied to problems where the use of Monte Carlo approximations is computationally infeasible

    Gibbs optimal design of experiments

    Full text link
    Bayesian optimal design of experiments is a well-established approach to planning experiments. Briefly, a probability distribution, known as a statistical model, for the responses is assumed which is dependent on a vector of unknown parameters. A utility function is then specified which gives the gain in information for estimating the true value of the parameters using the Bayesian posterior distribution. A Bayesian optimal design is given by maximising the expectation of the utility with respect to the joint distribution given by the statistical model and prior distribution for the true parameter values. The approach takes account of the experimental aim via specification of the utility and of all assumed sources of uncertainty via the expected utility. However, it is predicated on the specification of the statistical model. Recently, a new type of statistical inference, known as Gibbs (or General Bayesian) inference, has been advanced. This is Bayesian-like, in that uncertainty on unknown quantities is represented by a posterior distribution, but does not necessarily rely on specification of a statistical model. Thus the resulting inference should be less sensitive to misspecification of the statistical model. The purpose of this paper is to propose Gibbs optimal design: a framework for optimal design of experiments for Gibbs inference. The concept behind the framework is introduced along with a computational approach to find Gibbs optimal designs in practice. The framework is demonstrated on exemplars including linear models, and experiments with count and time-to-event responses

    Tropical cyclone contribution to extreme rainfall over southwest Pacific Island nations

    Get PDF
    Southwest Pacific nations are among some of the worst impacted and most vulnerable globally in terms of tropical cyclone (TC)-induced flooding and accompanying risks. This study objectively quantifies the fractional contribution of TCs to extreme rainfall (hereafter, TC contributions) in the context of climate variability and change. We show that TC contributions to extreme rainfall are substantially enhanced during active phases of the Madden–Julian Oscillation and by El Niño conditions (particularly over the eastern southwest Pacific region); this enhancement is primarily attributed to increased TC activity during these event periods. There are also indications of increasing intensities of TC-induced extreme rainfall events over the past few decades. A key part of this work involves development of sophisticated Bayesian regression models for individual island nations in order to better understand the synergistic relationships between TC-induced extreme rainfall and combinations of various climatic drivers that modulate the relationship. Such models are found to be very useful for not only assessing probabilities of TC- and non-TC induced extreme rainfall events but also evaluating probabilities of extreme rainfall for cases with different underlying climatic conditions. For example, TC-induced extreme rainfall probability over Samoa can vary from ~ 95 to ~ 75% during a La Niña period, if it coincides with an active or inactive phase of the MJO, and can be reduced to ~ 30% during a combination of El Niño period and inactive phase of the MJO. Several other such cases have been assessed for different island nations, providing information that have potentially important implications for planning and preparing for TC risks in vulnerable Pacific Island nations. © 2021, The Author(s). *Please note that there are multiple authors for this article therefore only the name of the first 5 including Federation University Australia affiliate “Anil Deo and Savin Chand” is provided in this record*

    A new approach to spatial data interpolation using higher-order statistics

    Get PDF
    Interpolation techniques for spatial data have been applied frequently in various fields of geosciences. Although most conventional interpolation methods assume that it is sufficient to use first- and second-order statistics to characterize random fields, researchers have now realized that these methods cannot always provide reliable interpolation results, since geological and environmental phenomena tend to be very complex, presenting non-Gaussian distribution and/or non-linear inter-variable relationship. This paper proposes a new approach to the interpolation of spatial data, which can be applied with great flexibility. Suitable cross-variable higher-order spatial statistics are developed to measure the spatial relationship between the random variable at an unsampled location and those in its neighbourhood. Given the computed cross-variable higher-order spatial statistics, the conditional probability density function is approximated via polynomial expansions, which is then utilized to determine the interpolated value at the unsampled location as an expectation. In addition, the uncertainty associated with the interpolation is quantified by constructing prediction intervals of interpolated values. The proposed method is applied to a mineral deposit dataset, and the results demonstrate that it outperforms kriging methods in uncertainty quantification. The introduction of the cross-variable higher-order spatial statistics noticeably improves the quality of the interpolation since it enriches the information that can be extracted from the observed data, and this benefit is substantial when working with data that are sparse or have non-trivial dependence structures
    • …
    corecore