10 research outputs found

    Estimating the evidence of selection and the reliability of inference in unigenic evolution

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Unigenic evolution is a large-scale mutagenesis experiment used to identify residues that are potentially important for protein function. Both currently-used methods for the analysis of unigenic evolution data analyze 'windows' of contiguous sites, a strategy that increases statistical power but incorrectly assumes that functionally-critical sites are contiguous. In addition, both methods require the questionable assumption of asymptotically-large sample size due to the presumption of approximate normality.</p> <p>Results</p> <p>We develop a novel approach, termed the Evidence of Selection (EoS), removing the assumption that functionally important sites are adjacent in sequence and and explicitly modelling the effects of limited sample-size. Precise statistical derivations show that the EoS score can be easily interpreted as an expected log-odds-ratio between two competing hypotheses, namely, the hypothetical presence or absence of functional selection for a given site. Using the EoS score, we then develop selection criteria by which functionally-important yet non-adjacent sites can be identified. An approximate power analysis is also developed to estimate the reliability of inference given the data. We validate and demonstrate the the practical utility of our method by analysis of the homing endonuclease <monospace>I-Bmol</monospace>, comparing our predictions with the results of existing methods.</p> <p>Conclusions</p> <p>Our method is able to assess both the evidence of selection at individual amino acid sites and estimate the reliability of those inferences. Experimental validation with <monospace>I-Bmol</monospace> proves its utility to identify functionally-important residues of poorly characterized proteins, demonstrating increased sensitivity over previous methods without loss of specificity. With the ability to guide the selection of precise experimental mutagenesis conditions, our method helps make unigenic analysis a more broadly applicable technique with which to probe protein function.</p> <p>Availability</p> <p>Software to compute, plot, and summarize EoS data is available as an open-source package called 'unigenic' for the 'R' programming language at <url>http://www.fernandes.org/txp/article/13/an-analytical-framework-for-unigenic-evolution</url>.</p

    Assessing the Non-Uniqueness of a Well Test Interpretation Model Using a Bayesian Approach

    No full text
    Objectives/Scope: A stable, single-well deconvolution algorithm has been introduced for well test analysis in the early 2000’s, that allows to obtain information about the reservoir system not always available from individual flow periods, for example the presence of heterogeneities and boundaries. One issue, recognised but largely ignored, is that of uncertainty in well test analysis results and non-uniqueness of the interpretation model. In a previous paper (SPE 164870), we assessed these with a Monte Carlo approach, where multiple deconvolutions were performed over the ranges of expected uncertainties affecting the data (Monte Carlo deconvolution). Methods, Procedures, Process: In this paper, we use a non-linear Bayesian regression model based on models of reservoir behaviour in order to make inferences about the interpretation model. This allows us to include uncertainty for the measurements which are usually contaminated with large observational errors. We combine the likelihood with flexible probability distributions for the inputs (priors), and we use Markov Chain Monte Carlo algorithms in order to approximate the probability distribution of the result (posterior). Results, Observations, Conclusions: We validate and illustrate the use of the algorithm by applying it to the same synthetic and field data sets as in SPE 164870, using a variety of tools to summarise and visualise the posterior distribution, and to carry out model selection. Novel/Additive Information: The approach used in this paper has several advantages over Monte Carlo deconvolution: (1) it gives access to meaningful system parameters associated with the flow behaviour in the reservoir; (2) it makes it possible to incorporate prior knowledge in order to exclude non-physical results; and (3) it allows to quantify parameter uncertainty in a principled way by exploiting the advantages of the Bayesian approach

    A Stochastic Grammar for Natural Shapes

    No full text
    We consider object detection using a generic model for natural shapes. A common approach for object recognition involves matching object models directly to images. Another approach involves building intermediate representations via a generic grouping processes. We argue that these two processes (model-based recognition and grouping) may use similar computational mechanisms. By defining a generic model for shapes we can use model-based techniques to implement a mid-level vision grouping process.

    RGB-D image-based Object Detection: From traditional methods to deep learning techniques

    No full text
    Object detection from RGB images is a long-standing problem in image processing and computer vision. It has applications in various domains including robotics, surveillance, human-computer interaction, and medical diagnosis. With the availability of low cost 3D scanners, a large number of RGB-D object detection approaches have been proposed in the past years. This chapter provides a comprehensive survey of the recent developments in this field. We structure the chapter into two parts; the focus of the first part is on techniques that are based on hand-crafted features combined with machine learning algorithms. The focus of the second part is on the more recent work, which is based on deep learning. Deep learning techniques, coupled with the availability of large training datasets, have now revolutionized the field of computer vision, including RGB-D object detection, achieving an unprecedented level of performance. We survey the key contributions, summarize the most commonly used pipelines, discuss their benefits and limitations, and highlight some important directions for future research

    Elastic shape analysis of boundaries of planar objects with multiple components and arbitrary topologies

    No full text
    We consider boundaries of planar objects as level set distance functions and present a Riemannian metric for their comparison and analysis. The metric is based on a parameterization-invariant framework for shape analysis of quadrilateral surfaces. Most previous Riemannian formulations of 2D shape analysis are restricted to curves that can be parameterized with a single parameter domain. However, 2D shapes may contain multiple connected components and many internal details that cannot be captured with such parameterizations. In this paper we propose to register planar curves of arbitrary topologies by utilizing the re-parameterization group of quadrilateral surfaces. The criterion used for computing this registration is a proper distance, which can be used to quantify differences between the level set functions and is especially useful in classification. We demonstrate this framework with multiple examples using toy curves, medical imaging data, subsets of the TOSCA data set, 2D hand-drawn sketches, and a 2D version of the SHREC07 data set. We demonstrate that our method outperforms the state-of-the-art in the classification of 2D sketches and performs well compared to other state-of-the-art methods on complex planar shapes
    corecore