136 research outputs found
Context-Specific Independence in Bayesian Networks
Bayesian networks provide a language for qualitatively representing the
conditional independence properties of a distribution. This allows a natural
and compact representation of the distribution, eases knowledge acquisition,
and supports effective inference algorithms. It is well-known, however, that
there are certain independencies that we cannot capture qualitatively within
the Bayesian network structure: independencies that hold only in certain
contexts, i.e., given a specific assignment of values to certain variables. In
this paper, we propose a formal notion of context-specific independence (CSI),
based on regularities in the conditional probability tables (CPTs) at a node.
We present a technique, analogous to (and based on) d-separation, for
determining when such independence holds in a given network. We then focus on a
particular qualitative representation scheme - tree-structured CPTs - for
capturing CSI. We suggest ways in which this representation can be used to
support effective inference algorithms. In particular, we present a structural
decomposition of the resulting network which can improve the performance of
clustering algorithms, and an alternative algorithm based on cutset
conditioning.Comment: Appears in Proceedings of the Twelfth Conference on Uncertainty in
Artificial Intelligence (UAI1996
Markov random fields factorization with context-specific independences
Markov random fields provide a compact representation of joint probability
distributions by representing its independence properties in an undirected
graph. The well-known Hammersley-Clifford theorem uses these conditional
independences to factorize a Gibbs distribution into a set of factors. However,
an important issue of using a graph to represent independences is that it
cannot encode some types of independence relations, such as the
context-specific independences (CSIs). They are a particular case of
conditional independences that is true only for a certain assignment of its
conditioning set; in contrast to conditional independences that must hold for
all its assignments. This work presents a method for factorizing a Markov
random field according to CSIs present in a distribution, and formally
guarantees that this factorization is correct. This is presented in our main
contribution, the context-specific Hammersley-Clifford theorem, a
generalization to CSIs of the Hammersley-Clifford theorem that applies for
conditional independences.Comment: 7 page
Independence of Causal Influence and Clique Tree Propagation
This paper explores the role of independence of causal influence (ICI) in
Bayesian network inference. ICI allows one to factorize a conditional
probability table into smaller pieces. We describe a method for exploiting the
factorization in clique tree propagation (CTP) - the state-of-the-art exact
inference algorithm for Bayesian networks. We also present empirical results
showing that the resulting algorithm is significantly more efficient than the
combination of CTP and previous techniques for exploiting ICI.Comment: Appears in Proceedings of the Thirteenth Conference on Uncertainty in
Artificial Intelligence (UAI1997
Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks
Particle filters (PFs) are powerful sampling-based inference/learning
algorithms for dynamic Bayesian networks (DBNs). They allow us to treat, in a
principled way, any type of probability distribution, nonlinearity and
non-stationarity. They have appeared in several fields under such names as
"condensation", "sequential Monte Carlo" and "survival of the fittest". In this
paper, we show how we can exploit the structure of the DBN to increase the
efficiency of particle filtering, using a technique known as
Rao-Blackwellisation. Essentially, this samples some of the variables, and
marginalizes out the rest exactly, using the Kalman filter, HMM filter,
junction tree algorithm, or any other finite dimensional optimal filter. We
show that Rao-Blackwellised particle filters (RBPFs) lead to more accurate
estimates than standard PFs. We demonstrate RBPFs on two problems, namely
non-stationary online regression with radial basis function networks and robot
localization and map building. We also discuss other potential application
areas and provide references to some finite dimensional optimal filters.Comment: Appears in Proceedings of the Sixteenth Conference on Uncertainty in
Artificial Intelligence (UAI2000
Sum-Product Network Decompilation
There exists a dichotomy between classical probabilistic graphical models,
such as Bayesian networks (BNs), and modern tractable models, such as
sum-product networks (SPNs). The former generally have intractable inference,
but provide a high level of interpretability, while the latter admits a wide
range of tractable inference routines, but are typically harder to interpret.
Due to this dichotomy, tools to convert between BNs and SPNs are desirable.
While one direction -- compiling BNs into SPNs -- is well discussed in
Darwiche's seminal work on arithmetic circuit compilation, the converse
direction -- decompiling SPNs into BNs -- has received surprisingly little
attention.
In this paper, we fill this gap by proposing SPN2BN, an algorithm that
decompiles an SPN into a BN. SPN2BN has several salient features when compared
to the only other two works decompiling SPNs. Most significantly, the BNs
returned by SPN2BN are minimal independence-maps that are more parsimonious
with respect to the introduction of latent variables. Secondly, the output BN
produced by SPN2BN can be precisely characterized with respect to a compiled
BN. More specifically, a certain set of directed edges will be added to the
input BN, giving what we will call the moral-closure. Lastly, it is established
that our compilation-decompilation process is idempotent. This has practical
significance as it limits the size of the decompiled SPN
Efficient Inference in Large Discrete Domains
In this paper we examine the problem of inference in Bayesian Networks with
discrete random variables that have very large or even unbounded domains. For
example, in a domain where we are trying to identify a person, we may have
variables that have as domains, the set of all names, the set of all postal
codes, or the set of all credit card numbers. We cannot just have big tables of
the conditional probabilities, but need compact representations. We provide an
inference algorithm, based on variable elimination, for belief networks
containing both large domain and normal discrete random variables. We use
intensional (i.e., in terms of procedures) and extensional (in terms of listing
the elements) representations of conditional probabilities and of the
intermediate factors.Comment: Appears in Proceedings of the Nineteenth Conference on Uncertainty in
Artificial Intelligence (UAI2003
Toward General Analysis of Recursive Probability Models
There is increasing interest within the research community in the design and
use of recursive probability models. Although there still remains concern about
computational complexity costs and the fact that computing exact solutions can
be intractable for many nonrecursive models and impossible in the general case
for recursive problems, several research groups are actively developing
computational techniques for recursive stochastic languages. We have developed
an extension to the traditional lambda-calculus as a framework for families of
Turing complete stochastic languages. We have also developed a class of exact
inference algorithms based on the traditional reductions of the
lambda-calculus. We further propose that using the deBruijn notation (a
lambda-calculus notation with nameless dummies) supports effective caching in
such systems (caching being an essential component of efficient computation).
Finally, our extension to the lambda-calculus offers a foundation and general
theory for the construction of recursive stochastic modeling languages as well
as promise for effective caching and efficient approximation algorithms for
inference.Comment: Appears in Proceedings of the Seventeenth Conference on Uncertainty
in Artificial Intelligence (UAI2001
YGGDRASIL - A Statistical Package for Learning Split Models
There are two main objectives of this paper. The first is to present a
statistical framework for models with context specific independence structures,
i.e., conditional independences holding only for sepcific values of the
conditioning variables. This framework is constituted by the class of split
models. Split models are extension of graphical models for contigency tables
and allow for a more sophisticiated modelling than graphical models. The
treatment of split models include estimation, representation and a Markov
property for reading off those independencies holding in a specific context.
The second objective is to present a software package named YGGDRASIL which is
designed for statistical inference in split models, i.e., for learning such
models on the basis of data.Comment: Appears in Proceedings of the Sixteenth Conference on Uncertainty in
Artificial Intelligence (UAI2000
Context-Specific Approximation in Probabilistic Inference
There is evidence that the numbers in probabilistic inference don't really
matter. This paper considers the idea that we can make a probabilistic model
simpler by making fewer distinctions. Unfortunately, the level of a Bayesian
network seems too coarse; it is unlikely that a parent will make little
difference for all values of the other parents. In this paper we consider an
approximation scheme where distinctions can be ignored in some contexts, but
not in other contexts. We elaborate on a notion of a parent context that allows
a structured context-specific decomposition of a probability distribution and
the associated probabilistic inference scheme called probabilistic partial
evaluation (Poole 1997). This paper shows a way to simplify a probabilistic
model by ignoring distinctions which have similar probabilities, a method to
exploit the simpler model, a bound on the resulting errors, and some
preliminary empirical results on simple networks.Comment: Appears in Proceedings of the Fourteenth Conference on Uncertainty in
Artificial Intelligence (UAI1998
User Interface Tools for Navigation in Conditional Probability Tables and Elicitation of Probabilities in Bayesian Networks
Elicitation of probabilities is one of the most laborious tasks in building
decision-theoretic models, and one that has so far received only moderate
attention in decision-theoretic systems. We propose a set of user interface
tools for graphical probabilistic models, focusing on two aspects of
probability elicitation: (1) navigation through conditional probability tables
and (2) interactive graphical assessment of discrete probability distributions.
We propose two new graphical views that aid navigation in very large
conditional probability tables: the CPTree (Conditional Probability Tree) and
the SCPT (shrinkable Conditional Probability Table). Based on what is known
about graphical presentation of quantitative data to humans, we offer several
useful enhancements to probability wheel and bar graph, including different
chart styles and options that can be adapted to user preferences and needs. We
present the results of a simple usability study that proves the value of the
proposed tools.Comment: Appears in Proceedings of the Sixteenth Conference on Uncertainty in
Artificial Intelligence (UAI2000
- …