28,652 research outputs found

    Characterization and Learning of Causal Graphs with Small Conditioning Sets

    Full text link
    Constraint-based causal discovery algorithms learn part of the causal graph structure by systematically testing conditional independences observed in the data. These algorithms, such as the PC algorithm and its variants, rely on graphical characterizations of the so-called equivalence class of causal graphs proposed by Pearl. However, constraint-based causal discovery algorithms struggle when data is limited since conditional independence tests quickly lose their statistical power, especially when the conditioning set is large. To address this, we propose using conditional independence tests where the size of the conditioning set is upper bounded by some integer kk for robust causal discovery. The existing graphical characterizations of the equivalence classes of causal graphs are not applicable when we cannot leverage all the conditional independence statements. We first define the notion of kk-Markov equivalence: Two causal graphs are kk-Markov equivalent if they entail the same conditional independence constraints where the conditioning set size is upper bounded by kk. We propose a novel representation that allows us to graphically characterize kk-Markov equivalence between two causal graphs. We propose a sound constraint-based algorithm called the kk-PC algorithm for learning this equivalence class. Finally, we conduct synthetic, and semi-synthetic experiments to demonstrate that the kk-PC algorithm enables more robust causal discovery in the small sample regime compared to the baseline PC algorithm.Comment: 30 page

    An Upper Bound for Random Measurement Error in Causal Discovery

    Get PDF
    Causal discovery algorithms infer causal relations from data based on several assumptions, including notably the absence of measurement error. However, this assumption is most likely violated in practical applications, which may result in erroneous, irreproducible results. In this work we show how to obtain an upper bound for the variance of random measurement error from the covariance matrix of measured variables and how to use this upper bound as a correction for constraint-based causal discovery. We demonstrate a practical application of our approach on both simulated data and real-world protein signaling data.Comment: Published in Proceedings of the 34th Annual Conference on Uncertainty in Artificial Intelligence (UAI-18
    • …
    corecore