33,157 research outputs found

    Probabilistic Constraint Logic Programming

    Full text link
    This paper addresses two central problems for probabilistic processing models: parameter estimation from incomplete data and efficient retrieval of most probable analyses. These questions have been answered satisfactorily only for probabilistic regular and context-free models. We address these problems for a more expressive probabilistic constraint logic programming model. We present a log-linear probability model for probabilistic constraint logic programming. On top of this model we define an algorithm to estimate the parameters and to select the properties of log-linear models from incomplete data. This algorithm is an extension of the improved iterative scaling algorithm of Della-Pietra, Della-Pietra, and Lafferty (1995). Our algorithm applies to log-linear models in general and is accompanied with suitable approximation methods when applied to large data spaces. Furthermore, we present an approach for searching for most probable analyses of the probabilistic constraint logic programming model. This method can be applied to the ambiguity resolution problem in natural language processing applications.Comment: 35 pages, uses sfbart.cl

    What May Visualization Processes Optimize?

    Full text link
    In this paper, we present an abstract model of visualization and inference processes and describe an information-theoretic measure for optimizing such processes. In order to obtain such an abstraction, we first examined six classes of workflows in data analysis and visualization, and identified four levels of typical visualization components, namely disseminative, observational, analytical and model-developmental visualization. We noticed a common phenomenon at different levels of visualization, that is, the transformation of data spaces (referred to as alphabets) usually corresponds to the reduction of maximal entropy along a workflow. Based on this observation, we establish an information-theoretic measure of cost-benefit ratio that may be used as a cost function for optimizing a data visualization process. To demonstrate the validity of this measure, we examined a number of successful visualization processes in the literature, and showed that the information-theoretic measure can mathematically explain the advantages of such processes over possible alternatives.Comment: 10 page

    Evaluation of IoT-Based Computational Intelligence Tools for DNA Sequence Analysis in Bioinformatics

    Full text link
    In contemporary age, Computational Intelligence (CI) performs an essential role in the interpretation of big biological data considering that it could provide all of the molecular biology and DNA sequencing computations. For this purpose, many researchers have attempted to implement different tools in this field and have competed aggressively. Hence, determining the best of them among the enormous number of available tools is not an easy task, selecting the one which accomplishes big data in the concise time and with no error can significantly improve the scientist's contribution in the bioinformatics field. This study uses different analysis and methods such as Fuzzy, Dempster-Shafer, Murphy and Entropy Shannon to provide the most significant and reliable evaluation of IoT-based computational intelligence tools for DNA sequence analysis. The outcomes of this study can be advantageous to the bioinformatics community, researchers and experts in big biological data
    • …
    corecore