5,028 research outputs found
Lower Complexity Bounds for Lifted Inference
One of the big challenges in the development of probabilistic relational (or
probabilistic logical) modeling and learning frameworks is the design of
inference techniques that operate on the level of the abstract model
representation language, rather than on the level of ground, propositional
instances of the model. Numerous approaches for such "lifted inference"
techniques have been proposed. While it has been demonstrated that these
techniques will lead to significantly more efficient inference on some specific
models, there are only very recent and still quite restricted results that show
the feasibility of lifted inference on certain syntactically defined classes of
models. Lower complexity bounds that imply some limitations for the feasibility
of lifted inference on more expressive model classes were established early on
in (Jaeger 2000). However, it is not immediate that these results also apply to
the type of modeling languages that currently receive the most attention, i.e.,
weighted, quantifier-free formulas. In this paper we extend these earlier
results, and show that under the assumption that NETIME =/= ETIME, there is no
polynomial lifted inference algorithm for knowledge bases of weighted,
quantifier- and function-free formulas. Further strengthening earlier results,
this is also shown to hold for approximate inference, and for knowledge bases
not containing the equality predicate.Comment: To appear in Theory and Practice of Logic Programming (TPLP
Heuristic Ranking in Tightly Coupled Probabilistic Description Logics
The Semantic Web effort has steadily been gaining traction in the recent
years. In particular,Web search companies are recently realizing that their
products need to evolve towards having richer semantic search capabilities.
Description logics (DLs) have been adopted as the formal underpinnings for
Semantic Web languages used in describing ontologies. Reasoning under
uncertainty has recently taken a leading role in this arena, given the nature
of data found on theWeb. In this paper, we present a probabilistic extension of
the DL EL++ (which underlies the OWL2 EL profile) using Markov logic networks
(MLNs) as probabilistic semantics. This extension is tightly coupled, meaning
that probabilistic annotations in formulas can refer to objects in the
ontology. We show that, even though the tightly coupled nature of our language
means that many basic operations are data-intractable, we can leverage a
sublanguage of MLNs that allows to rank the atomic consequences of an ontology
relative to their probability values (called ranking queries) even when these
values are not fully computed. We present an anytime algorithm to answer
ranking queries, and provide an upper bound on the error that it incurs, as
well as a criterion to decide when results are guaranteed to be correct.Comment: Appears in Proceedings of the Twenty-Eighth Conference on Uncertainty
in Artificial Intelligence (UAI2012
Towards Log-Linear Logics with Concrete Domains
We present (M denotes Markov logic networks) an
extension of the log-linear description logics -LL with
concrete domains, nominals, and instances. We use Markov logic networks (MLNs)
in order to find the most probable, classified and coherent
ontology from an knowledge base. In particular, we develop
a novel way to deal with concrete domains (also known as datatypes) by
extending MLN's cutting plane inference (CPI) algorithm.Comment: StarAI201
Structurally Tractable Uncertain Data
Many data management applications must deal with data which is uncertain,
incomplete, or noisy. However, on existing uncertain data representations, we
cannot tractably perform the important query evaluation tasks of determining
query possibility, certainty, or probability: these problems are hard on
arbitrary uncertain input instances. We thus ask whether we could restrict the
structure of uncertain data so as to guarantee the tractability of exact query
evaluation. We present our tractability results for tree and tree-like
uncertain data, and a vision for probabilistic rule reasoning. We also study
uncertainty about order, proposing a suitable representation, and study
uncertain data conditioned by additional observations.Comment: 11 pages, 1 figure, 1 table. To appear in SIGMOD/PODS PhD Symposium
201
- …