312 research outputs found

    Letter to the editor

    Get PDF

    Minimum phi-divergence estimator and hierarchical testing in loglinear models

    Get PDF
    In this paper we consider inference based on very general divergence measures, under assumptions of multinomial sampling and loglinear models. We define the minimum phi-divergence estimator, which is seen to be a generalization of the maximum likelihood estimator. This estimator is then used in a phi-divergence goodness-of-fit statistic, which is the basis of two new statistics for solving the problem of testing a nested sequence of loglinear models

    Random set theory and problems of modeling

    Get PDF
    The three- or four-dimensional world in which we live is full of objects to be measured and summarized. Very often a parsimonious finite collection of measurements is enough for scientific investigation into an object’s genesis and evolution. There is a growing need, however, to describe and model objects through their form as well as their size. The purpose of this article is to show the potentials and limitations of a probabilistic and statistical approach. Collections of objects (the data) are assimilated to a random set (the model), whose parameters provide description and/or explanation

    Mine boundary detection using partially ordered Markov models

    Get PDF
    Detection of objects in images in an automated fashion is necessary for many applications, including automated target recognition. In this paper, we present results of an automated boundary detection procedure using a new subclass of Markov random fields (MRFs), called partially ordered Markov models (POMMs). POMMs offer computational advantages over general MRFs. We show how a POMM can model the boundaries in an image. Our algorithm for boundary detection uses a Bayesian approach to build a posterior boundary model that locates edges of objects having a closed loop boundary. We apply our method to images of mines with very good results. 2004 Copyright SPIE - The International Society for Optical Engineering

    Class-based probability estimation using a semantic hierarchy

    Get PDF
    This article concerns the estimation of a particular kind of probability, namely, the probability of a noun sense appearing as a particular argument of a predicate. In order to overcome the accompanying sparse-data problem, the proposal here is to define the probabilities in terms of senses from a semantic hierarchy and exploit the fact that the senses can be grouped into classes consisting of semantically similar senses. There is a particular focus on the problem of how to determine a suitable class for a given sense, or, alternatively, how to determine a suitable level of generalization in the hierarchy. A procedure is developed that uses a chi-square test to determine a suitable level of generalization. In order to test the performance of the estimation method, a pseudo-disambiguation task is used, together with two alternative estimation methods. Each method uses a different generalization procedure; the first alternative uses the minimum description length principle, and the second uses Resnik's measure of selectional preference. In addition, the performance of our method is investigated using both the standard Pearson chi-square statistic and the log-likelihood chi-square statistic

    Accounting for Uncertainty in Ecological Analysis: The Strengths and Limitations of Hierarchical Statistical Modeling

    Get PDF
    Copyright by the Ecological Society of America.Analyses of ecological data should account for the uncertainty in the process(es) that generated the data. However, accounting for these uncertainties is a difficult task, since ecology is known for its complexity. Measurement and/or process errors are often the only sources of uncertainty modeled when addressing complex ecological problems, yet analyses should also account for uncertainty in sampling design, in model specification, in parameters governing the specified model, and in initial and boundary conditions. Only then can we be confident in the scientific inferences and forecasts made from an analysis. Probability and statistics provide a framework that accounts for multiple sources of uncertainty. Given the complexities of ecological studies, the hierarchical statistical model is an invaluable tool. This approach is not new in ecology, and there are many examples (both Bayesian and non-Bayesian) in the literature illustrating the benefits of this approach. In this article, we provide a baseline for concepts, notation, and methods, from which discussion on hierarchical statistical modeling in ecology can proceed. We have also planted some seeds for discussion and tried to show where the practical difficulties lie. Our thesis is that hierarchical statistical modeling is a powerful way of approaching ecological analysis in the presence of inevitable but quantifiable uncertainties, even if practical issues sometimes require pragmatic compromises

    Accounting for uncertainty in ecological analysis: the strengths and limitations of hierarchical statistical modeling

    Get PDF
    Analyses of ecological data should account for the uncertainty in the process(es) that generated the data. However, accounting for these uncertainties is a difficult task, since ecology is known for its complexity. Measurement and/or process errors are often the only sources of uncertainty modeled when addressing complex ecological problems, yet analyses should also account for uncertainty in sampling design, in model specification, in parameters governing the specified model, and in initial and boundary conditions. Only then can we be confident in the scientific inferences and forecasts made from an analysis. Probability and statistics provide a framework that accounts for multiple sources of uncertainty. Given the complexities of ecological studies, the hierarchical statistical model is an invaluable tool. This approach is not new in ecology, and there are many examples (both Bayesian and non-Bayesian) in the literature illustrating the benefits of this approach. In this article, we provide a baseline for concepts, notation, and methods, from which discussion on hierarchical statistical modeling in ecology can proceed. We have also planted some seeds for discussion and tried to show where the practical difficulties lie. Our thesis is that hierarchical statistical modeling is a powerful way of approaching ecological analysis in the presence of inevitable but quantifiable uncertainties, even if practical issues sometimes require pragmatic compromises

    Bayesian optimization for materials design

    Full text link
    We introduce Bayesian optimization, a technique developed for optimizing time-consuming engineering simulations and for fitting machine learning models on large datasets. Bayesian optimization guides the choice of experiments during materials design and discovery to find good material designs in as few experiments as possible. We focus on the case when materials designs are parameterized by a low-dimensional vector. Bayesian optimization is built on a statistical technique called Gaussian process regression, which allows predicting the performance of a new design based on previously tested designs. After providing a detailed introduction to Gaussian process regression, we introduce two Bayesian optimization methods: expected improvement, for design problems with noise-free evaluations; and the knowledge-gradient method, which generalizes expected improvement and may be used in design problems with noisy evaluations. Both methods are derived using a value-of-information analysis, and enjoy one-step Bayes-optimality

    National CO2 budgets (2015–2020) inferred from atmospheric CO2 observations in support of the Global Stocktake

    Get PDF
    Accurate accounting of emissions and removals of CO2 is critical for the planning and verification of emission reduction targets in support of the Paris Agreement. Here, we present a pilot dataset of country-specific net carbon exchange (NCE; fossil plus terrestrial ecosystem fluxes) and terrestrial carbon stock changes aimed at informing countries’ carbon budgets. These estimates are based on "top-down" NCE outputs from the v10 Orbiting Carbon Observatory (OCO-2) modeling intercomparison project (MIP), wherein an ensemble of inverse modeling groups conducted standardized experiments assimilating OCO-2 column-averaged dry-air mole fraction (XCO2) retrievals (ACOS v10), in situ CO2 measurements, or combinations of these data. The v10 OCO-2 MIP NCE estimates are combined with "bottom-up" estimates of fossil fuel emissions and lateral carbon fluxes to estimate changes in terrestrial carbon stocks, which are impacted by anthropogenic and natural drivers. These flux and stock change estimates are reported annually (2015–2020) as both a global 1° × 1° gridded dataset and as a country-level dataset. Across the v10 OCO-2 MIP experiments, we obtain increases in the ensemble median terrestrial carbon stocks of 3.29–4.58 PgCO2 yr-1 (0.90–1.25 PgC yr-1). This is a result of broad increases in terrestrial carbon stocks across the northern extratropics, while the tropics generally have stock losses but with considerable regional variability and differences between v10 OCO-2 MIP experiments. We discuss the state of the science for tracking emissions and removals using top-down methods, including current limitations and future developments towards top-down monitoring and verification systems
    • …
    corecore