9 research outputs found
Smooth, identifiable supermodels of discrete DAG models with latent variables
We provide a parameterization of the discrete nested Markov model, which is a
supermodel that approximates DAG models (Bayesian network models) with latent
variables. Such models are widely used in causal inference and machine
learning. We explicitly evaluate their dimension, show that they are curved
exponential families of distributions, and fit them to data. The
parameterization avoids the irregularities and unidentifiability of latent
variable models. The parameters used are all fully identifiable and
causally-interpretable quantities.Comment: 30 page
Graphs for margins of Bayesian networks
Directed acyclic graph (DAG) models, also called Bayesian networks, impose
conditional independence constraints on a multivariate probability
distribution, and are widely used in probabilistic reasoning, machine learning
and causal inference. If latent variables are included in such a model, then
the set of possible marginal distributions over the remaining (observed)
variables is generally complex, and not represented by any DAG. Larger classes
of mixed graphical models, which use multiple edge types, have been introduced
to overcome this; however, these classes do not represent all the models which
can arise as margins of DAGs. In this paper we show that this is because
ordinary mixed graphs are fundamentally insufficiently rich to capture the
variety of marginal models.
We introduce a new class of hyper-graphs, called mDAGs, and a latent
projection operation to obtain an mDAG from the margin of a DAG. We show that
each distinct marginal of a DAG model is represented by at least one mDAG, and
provide graphical results towards characterizing when two such marginal models
are the same. Finally we show that mDAGs correctly capture the marginal
structure of causally-interpreted DAGs under interventions on the observed
variables
Constraint-based Causal Discovery for Non-Linear Structural Causal Models with Cycles and Latent Confounders
We address the problem of causal discovery from data, making use of the
recently proposed causal modeling framework of modular structural causal models
(mSCM) to handle cycles, latent confounders and non-linearities. We introduce
{\sigma}-connection graphs ({\sigma}-CG), a new class of mixed graphs
(containing undirected, bidirected and directed edges) with additional
structure, and extend the concept of {\sigma}-separation, the appropriate
generalization of the well-known notion of d-separation in this setting, to
apply to {\sigma}-CGs. We prove the closedness of {\sigma}-separation under
marginalisation and conditioning and exploit this to implement a test of
{\sigma}-separation on a {\sigma}-CG. This then leads us to the first causal
discovery algorithm that can handle non-linear functional relations, latent
confounders, cyclic causal relationships, and data from different (stochastic)
perfect interventions. As a proof of concept, we show on synthetic data how
well the algorithm recovers features of the causal graph of modular structural
causal models.Comment: Accepted for publication in Conference on Uncertainty in Artificial
Intelligence 201
Marginal log-linear parameters for graphical Markov models
Marginal log-linear (MLL) models provide a flexible approach to multivariate
discrete data. MLL parametrizations under linear constraints induce a wide
variety of models, including models defined by conditional independences. We
introduce a sub-class of MLL models which correspond to Acyclic Directed Mixed
Graphs (ADMGs) under the usual global Markov property. We characterize for
precisely which graphs the resulting parametrization is variation independent.
The MLL approach provides the first description of ADMG models in terms of a
minimal list of constraints. The parametrization is also easily adapted to
sparse modelling techniques, which we illustrate using several examples of real
data.Comment: 36 page