1,328 research outputs found
Encoding Markov Logic Networks in Possibilistic Logic
Markov logic uses weighted formulas to compactly encode a probability
distribution over possible worlds. Despite the use of logical formulas, Markov
logic networks (MLNs) can be difficult to interpret, due to the often
counter-intuitive meaning of their weights. To address this issue, we propose a
method to construct a possibilistic logic theory that exactly captures what can
be derived from a given MLN using maximum a posteriori (MAP) inference.
Unfortunately, the size of this theory is exponential in general. We therefore
also propose two methods which can derive compact theories that still capture
MAP inference, but only for specific types of evidence. These theories can be
used, among others, to make explicit the hidden assumptions underlying an MLN
or to explain the predictions it makes.Comment: Extended version of a paper appearing in UAI 201
Induction of Interpretable Possibilistic Logic Theories from Relational Data
The field of Statistical Relational Learning (SRL) is concerned with learning
probabilistic models from relational data. Learned SRL models are typically
represented using some kind of weighted logical formulas, which make them
considerably more interpretable than those obtained by e.g. neural networks. In
practice, however, these models are often still difficult to interpret
correctly, as they can contain many formulas that interact in non-trivial ways
and weights do not always have an intuitive meaning. To address this, we
propose a new SRL method which uses possibilistic logic to encode relational
models. Learned models are then essentially stratified classical theories,
which explicitly encode what can be derived with a given level of certainty.
Compared to Markov Logic Networks (MLNs), our method is faster and produces
considerably more interpretable models.Comment: Longer version of a paper appearing in IJCAI 201
The PITA System: Tabling and Answer Subsumption for Reasoning under Uncertainty
Many real world domains require the representation of a measure of
uncertainty. The most common such representation is probability, and the
combination of probability with logic programs has given rise to the field of
Probabilistic Logic Programming (PLP), leading to languages such as the
Independent Choice Logic, Logic Programs with Annotated Disjunctions (LPADs),
Problog, PRISM and others. These languages share a similar distribution
semantics, and methods have been devised to translate programs between these
languages. The complexity of computing the probability of queries to these
general PLP programs is very high due to the need to combine the probabilities
of explanations that may not be exclusive. As one alternative, the PRISM system
reduces the complexity of query answering by restricting the form of programs
it can evaluate. As an entirely different alternative, Possibilistic Logic
Programs adopt a simpler metric of uncertainty than probability. Each of these
approaches -- general PLP, restricted PLP, and Possibilistic Logic Programming
-- can be useful in different domains depending on the form of uncertainty to
be represented, on the form of programs needed to model problems, and on the
scale of the problems to be solved. In this paper, we show how the PITA system,
which originally supported the general PLP language of LPADs, can also
efficiently support restricted PLP and Possibilistic Logic Programs. PITA
relies on tabling with answer subsumption and consists of a transformation
along with an API for library functions that interface with answer subsumption
Belief Revision with Uncertain Inputs in the Possibilistic Setting
This paper discusses belief revision under uncertain inputs in the framework
of possibility theory. Revision can be based on two possible definitions of the
conditioning operation, one based on min operator which requires a purely
ordinal scale only, and another based on product, for which a richer structure
is needed, and which is a particular case of Dempster's rule of conditioning.
Besides, revision under uncertain inputs can be understood in two different
ways depending on whether the input is viewed, or not, as a constraint to
enforce. Moreover, it is shown that M.A. Williams' transmutations, originally
defined in the setting of Spohn's functions, can be captured in this framework,
as well as Boutilier's natural revision.Comment: Appears in Proceedings of the Twelfth Conference on Uncertainty in
Artificial Intelligence (UAI1996
An Ordinal View of Independence with Application to Plausible Reasoning
An ordinal view of independence is studied in the framework of possibility
theory. We investigate three possible definitions of dependence, of increasing
strength. One of them is the counterpart to the multiplication law in
probability theory, and the two others are based on the notion of conditional
possibility. These two have enough expressive power to support the whole
possibility theory, and a complete axiomatization is provided for the strongest
one. Moreover we show that weak independence is well-suited to the problems of
belief change and plausible reasoning, especially to address the problem of
blocking of property inheritance in exception-tolerant taxonomic reasoning.Comment: Appears in Proceedings of the Tenth Conference on Uncertainty in
Artificial Intelligence (UAI1994
- …