33,157 research outputs found
Probabilistic Constraint Logic Programming
This paper addresses two central problems for probabilistic processing
models: parameter estimation from incomplete data and efficient retrieval of
most probable analyses. These questions have been answered satisfactorily only
for probabilistic regular and context-free models. We address these problems
for a more expressive probabilistic constraint logic programming model. We
present a log-linear probability model for probabilistic constraint logic
programming. On top of this model we define an algorithm to estimate the
parameters and to select the properties of log-linear models from incomplete
data. This algorithm is an extension of the improved iterative scaling
algorithm of Della-Pietra, Della-Pietra, and Lafferty (1995). Our algorithm
applies to log-linear models in general and is accompanied with suitable
approximation methods when applied to large data spaces. Furthermore, we
present an approach for searching for most probable analyses of the
probabilistic constraint logic programming model. This method can be applied to
the ambiguity resolution problem in natural language processing applications.Comment: 35 pages, uses sfbart.cl
What May Visualization Processes Optimize?
In this paper, we present an abstract model of visualization and inference
processes and describe an information-theoretic measure for optimizing such
processes. In order to obtain such an abstraction, we first examined six
classes of workflows in data analysis and visualization, and identified four
levels of typical visualization components, namely disseminative,
observational, analytical and model-developmental visualization. We noticed a
common phenomenon at different levels of visualization, that is, the
transformation of data spaces (referred to as alphabets) usually corresponds to
the reduction of maximal entropy along a workflow. Based on this observation,
we establish an information-theoretic measure of cost-benefit ratio that may be
used as a cost function for optimizing a data visualization process. To
demonstrate the validity of this measure, we examined a number of successful
visualization processes in the literature, and showed that the
information-theoretic measure can mathematically explain the advantages of such
processes over possible alternatives.Comment: 10 page
Evaluation of IoT-Based Computational Intelligence Tools for DNA Sequence Analysis in Bioinformatics
In contemporary age, Computational Intelligence (CI) performs an essential
role in the interpretation of big biological data considering that it could
provide all of the molecular biology and DNA sequencing computations. For this
purpose, many researchers have attempted to implement different tools in this
field and have competed aggressively. Hence, determining the best of them among
the enormous number of available tools is not an easy task, selecting the one
which accomplishes big data in the concise time and with no error can
significantly improve the scientist's contribution in the bioinformatics field.
This study uses different analysis and methods such as Fuzzy, Dempster-Shafer,
Murphy and Entropy Shannon to provide the most significant and reliable
evaluation of IoT-based computational intelligence tools for DNA sequence
analysis. The outcomes of this study can be advantageous to the bioinformatics
community, researchers and experts in big biological data
- …