92 research outputs found
Direct Cause
An interventionist account of causation characterizes causal relations in terms of changes resulting from particular interventions. We provide an example of a causal relation for which there does not exist an intervention satisfying the common interventionist standard. We consider adaptations that would save this standard and describe their implications for an interventionist account of causation. No adaptation preserves all the aspects that make the interventionist account appealing
Visual Causal Feature Learning
We provide a rigorous definition of the visual cause of a behavior that is
broadly applicable to the visually driven behavior in humans, animals, neurons,
robots and other perceiving systems. Our framework generalizes standard
accounts of causal learning to settings in which the causal variables need to
be constructed from micro-variables. We prove the Causal Coarsening Theorem,
which allows us to gain causal knowledge from observational data with minimal
experimental effort. The theorem provides a connection to standard inference
techniques in machine learning that identify features of an image that
correlate with, but may not cause, the target behavior. Finally, we propose an
active learning scheme to learn a manipulator function that performs optimal
manipulations on the image to automatically identify the visual cause of a
target behavior. We illustrate our inference and learning algorithms in
experiments based on both synthetic and real data.Comment: Accepted at UAI 201
Direct Cause
An interventionist account of causation characterizes causal relations in terms of changes resulting from particular interventions. We provide an example of a causal relation for which there does not exist an intervention satisfying the common interventionist standard. We consider adaptations that would save this standard and describe their implications for an interventionist account of causation. No adaptation preserves all the aspects that make the interventionist account appealing
Approximate Causal Abstraction
Scientific models describe natural phenomena at different levels of abstraction. Abstract descriptions can provide the basis for interventions on the system and explanation of observed phenomena at a level of granularity that is coarser than the most fundamental account of the system. Beckers and Halpern (2019), building on work of Rubenstein et al. (2017), developed an account of abstraction for causal models that is exact. Here we extend this account to the more realistic case where an abstract causal model offers only an approximation of the underlying system. We show how the resulting account handles the discrepancy that can arise between low- and high-level causal models of the same system, and in the process provide an account of how one causal model approximates another, a topic of independent interest. Finally, we extend the account of approximate abstractions to probabilistic causal models, indicating how and where uncertainty can enter into an approximate abstraction
Estimating Causal Direction and Confounding of Two Discrete Variables
We propose a method to classify the causal relationship between two discrete variables given only the joint distribution of the variables, acknowledging that the method is subject to an inherent baseline error. We assume that the causal system is acyclicity, but we do allow for hidden common causes. Our algorithm presupposes that the probability distributions P(C) of a cause C is independent from the probability distribution P(E∣C) of the cause-effect mechanism. While our classifier is trained with a Bayesian assumption of flat hyperpriors, we do not make this assumption about our test data. This work connects to recent developments on the identifiability of causal models over continuous variables under the assumption of "independent mechanisms". Carefully-commented Python notebooks that reproduce all our experiments are available online at http://vision.caltech.edu/~kchalupk/code.html
Fast Conditional Independence Test for Vector Variables with Large Sample Sizes
We present and evaluate the Fast (conditional) Independence Test (FIT) -- a
nonparametric conditional independence test. The test is based on the idea that
when , is not useful as a feature to predict
, as long as is also a regressor. On the contrary, if , might improve prediction results. FIT applies to
thousand-dimensional random variables with a hundred thousand samples in a
fraction of the time required by alternative methods. We provide an extensive
evaluation that compares FIT to six extant nonparametric independence tests.
The evaluation shows that FIT has low probability of making both Type I and
Type II errors compared to other tests, especially as the number of available
samples grows. Our implementation of FIT is publicly available
On the Number of Experiments Sufficient and in the Worst Case Necessary to Identify All Causal Relations Among N Variables
We show that if any number of variables are allowed to be simultaneously and
independently randomized in any one experiment, log2(N) + 1 experiments are
sufficient and in the worst case necessary to determine the causal relations
among N >= 2 variables when no latent variables, no sample selection bias and
no feedback cycles are present. For all K, 0 < K < 1/(2N) we provide an upper
bound on the number experiments required to determine causal structure when
each experiment simultaneously randomizes K variables. For large N, these
bounds are significantly lower than the N - 1 bound required when each
experiment randomizes at most one variable. For kmax < N/2, we show that
(N/kmax-1)+N/(2kmax)log2(kmax) experiments aresufficient and in the worst case
necessary. We over a conjecture as to the minimal number of experiments that
are in the worst case sufficient to identify all causal relations among N
observed variables that are a subset of the vertices of a DAG.Comment: Appears in Proceedings of the Twenty-First Conference on Uncertainty
in Artificial Intelligence (UAI2005
- …