85 research outputs found
Embedding Non-Ground Logic Programs into Autoepistemic Logic for Knowledge Base Combination
In the context of the Semantic Web, several approaches to the combination of
ontologies, given in terms of theories of classical first-order logic and rule
bases, have been proposed. They either cast rules into classical logic or limit
the interaction between rules and ontologies. Autoepistemic logic (AEL) is an
attractive formalism which allows to overcome these limitations, by serving as
a uniform host language to embed ontologies and nonmonotonic logic programs
into it. For the latter, so far only the propositional setting has been
considered. In this paper, we present three embeddings of normal and three
embeddings of disjunctive non-ground logic programs under the stable model
semantics into first-order AEL. While the embeddings all correspond with
respect to objective ground atoms, differences arise when considering
non-atomic formulas and combinations with first-order theories. We compare the
embeddings with respect to stable expansions and autoepistemic consequences,
considering the embeddings by themselves, as well as combinations with
classical theories. Our results reveal differences and correspondences of the
embeddings and provide useful guidance in the choice of a particular embedding
for knowledge combination.Comment: 52 pages, submitte
Complexity of Non-Monotonic Logics
Over the past few decades, non-monotonic reasoning has developed to be one of
the most important topics in computational logic and artificial intelligence.
Different ways to introduce non-monotonic aspects to classical logic have been
considered, e.g., extension with default rules, extension with modal belief
operators, or modification of the semantics. In this survey we consider a
logical formalism from each of the above possibilities, namely Reiter's default
logic, Moore's autoepistemic logic and McCarthy's circumscription.
Additionally, we consider abduction, where one is not interested in inferences
from a given knowledge base but in computing possible explanations for an
observation with respect to a given knowledge base.
Complexity results for different reasoning tasks for propositional variants
of these logics have been studied already in the nineties. In recent years,
however, a renewed interest in complexity issues can be observed. One current
focal approach is to consider parameterized problems and identify reasonable
parameters that allow for FPT algorithms. In another approach, the emphasis
lies on identifying fragments, i.e., restriction of the logical language, that
allow more efficient algorithms for the most important reasoning tasks. In this
survey we focus on this second aspect. We describe complexity results for
fragments of logical languages obtained by either restricting the allowed set
of operators (e.g., forbidding negations one might consider only monotone
formulae) or by considering only formulae in conjunctive normal form but with
generalized clause types.
The algorithmic problems we consider are suitable variants of satisfiability
and implication in each of the logics, but also counting problems, where one is
not only interested in the existence of certain objects (e.g., models of a
formula) but asks for their number.Comment: To appear in Bulletin of the EATC
Implementing Default and Autoepistemic Logics via the Logic of GK
The logic of knowledge and justified assumptions, also known as logic of
grounded knowledge (GK), was proposed by Lin and Shoham as a general logic for
nonmonotonic reasoning. To date, it has been used to embed in it default logic
(propositional case), autoepistemic logic, Turner's logic of universal
causation, and general logic programming under stable model semantics. Besides
showing the generality of GK as a logic for nonmonotonic reasoning, these
embeddings shed light on the relationships among these other logics. In this
paper, for the first time, we show how the logic of GK can be embedded into
disjunctive logic programming in a polynomial but non-modular translation with
new variables. The result can then be used to compute the extension/expansion
semantics of default logic, autoepistemic logic and Turner's logic of universal
causation by disjunctive ASP solvers such as claspD(-2), DLV, GNT and cmodels.Comment: Proceedings of the 15th International Workshop on Non-Monotonic
Reasoning (NMR 2014
Reasoning about Minimal Belief and Negation as Failure
We investigate the problem of reasoning in the propositional fragment of
MBNF, the logic of minimal belief and negation as failure introduced by
Lifschitz, which can be considered as a unifying framework for several
nonmonotonic formalisms, including default logic, autoepistemic logic,
circumscription, epistemic queries, and logic programming. We characterize the
complexity and provide algorithms for reasoning in propositional MBNF. In
particular, we show that entailment in propositional MBNF lies at the third
level of the polynomial hierarchy, hence it is harder than reasoning in all the
above mentioned propositional formalisms for nonmonotonic reasoning. We also
prove the exact correspondence between negation as failure in MBNF and negative
introspection in Moore's autoepistemic logic
The Complexity of Reasoning for Fragments of Autoepistemic Logic
Autoepistemic logic extends propositional logic by the modal operator L. A
formula that is preceded by an L is said to be "believed". The logic was
introduced by Moore 1985 for modeling an ideally rational agent's behavior and
reasoning about his own beliefs. In this paper we analyze all Boolean fragments
of autoepistemic logic with respect to the computational complexity of the
three most common decision problems expansion existence, brave reasoning and
cautious reasoning. As a second contribution we classify the computational
complexity of counting the number of stable expansions of a given knowledge
base. To the best of our knowledge this is the first paper analyzing the
counting problem for autoepistemic logic
Ultimate approximations in nonmonotonic knowledge representation systems
We study fixpoints of operators on lattices. To this end we introduce the
notion of an approximation of an operator. We order approximations by means of
a precision ordering. We show that each lattice operator O has a unique most
precise or ultimate approximation. We demonstrate that fixpoints of this
ultimate approximation provide useful insights into fixpoints of the operator
O.
We apply our theory to logic programming and introduce the ultimate
Kripke-Kleene, well-founded and stable semantics. We show that the ultimate
Kripke-Kleene and well-founded semantics are more precise then their standard
counterparts We argue that ultimate semantics for logic programming have
attractive epistemological properties and that, while in general they are
computationally more complex than the standard semantics, for many classes of
theories, their complexity is no worse.Comment: This paper was published in Principles of Knowledge Representation
and Reasoning, Proceedings of the Eighth International Conference (KR2002
Super Logic Programs
The Autoepistemic Logic of Knowledge and Belief (AELB) is a powerful
nonmonotic formalism introduced by Teodor Przymusinski in 1994. In this paper,
we specialize it to a class of theories called `super logic programs'. We argue
that these programs form a natural generalization of standard logic programs.
In particular, they allow disjunctions and default negation of arbibrary
positive objective formulas.
Our main results are two new and powerful characterizations of the static
semant ics of these programs, one syntactic, and one model-theoretic. The
syntactic fixed point characterization is much simpler than the fixed point
construction of the static semantics for arbitrary AELB theories. The
model-theoretic characterization via Kripke models allows one to construct
finite representations of the inherently infinite static expansions.
Both characterizations can be used as the basis of algorithms for query
answering under the static semantics. We describe a query-answering interpreter
for super programs which we developed based on the model-theoretic
characterization and which is available on the web.Comment: 47 pages, revised version of the paper submitted 10/200
- …