120 research outputs found

    Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models

    Full text link
    Shapley values underlie one of the most popular model-agnostic methods within explainable artificial intelligence. These values are designed to attribute the difference between a model's prediction and an average baseline to the different features used as input to the model. Being based on solid game-theoretic principles, Shapley values uniquely satisfy several desirable properties, which is why they are increasingly used to explain the predictions of possibly complex and highly non-linear machine learning models. Shapley values are well calibrated to a user's intuition when features are independent, but may lead to undesirable, counterintuitive explanations when the independence assumption is violated. In this paper, we propose a novel framework for computing Shapley values that generalizes recent work that aims to circumvent the independence assumption. By employing Pearl's do-calculus, we show how these 'causal' Shapley values can be derived for general causal graphs without sacrificing any of their desirable properties. Moreover, causal Shapley values enable us to separate the contribution of direct and indirect effects. We provide a practical implementation for computing causal Shapley values based on causal chain graphs when only partial information is available and illustrate their utility on a real-world example.Comment: Accepted at 34th Conference on Neural Information Processing Systems (NeurIPS 2020

    Learning Instantiation in First-Order Logic

    Get PDF
    Contains fulltext : 286055.pdf (Publisher’s version ) (Open Access)AITP 202

    Dynamical and Stationary Properties of On-line Learning from Finite Training Sets

    Full text link
    The dynamical and stationary properties of on-line learning from finite training sets are analysed using the cavity method. For large input dimensions, we derive equations for the macroscopic parameters, namely, the student-teacher correlation, the student-student autocorrelation and the learning force uctuation. This enables us to provide analytical solutions to Adaline learning as a benchmark. Theoretical predictions of training errors in transient and stationary states are obtained by a Monte Carlo sampling procedure. Generalization and training errors are found to agree with simulations. The physical origin of the critical learning rate is presented. Comparison with batch learning is discussed throughout the paper.Comment: 30 pages, 4 figure

    Investigation of topographical stability of the concave and convex Self-Organizing Map variant

    Get PDF
    We investigate, by a systematic numerical study, the parameter dependence of the stability of the Kohonen Self-Organizing Map and the Zheng and Greenleaf concave and convex learning with respect to different input distributions, input and output dimensions

    Effect of time-correlation of input patterns on the convergence of on-line learning

    Get PDF
    We studied the effects of time correlation of subsequent patterns on the convergence of on-line learning by a feedforward neural network with backpropagation algorithm. By using chaotic time series as sequences of correlated patterns, we found that the unexpected scaling of converging time with learning parameter emerges when time-correlated patterns accelerate learning process.Comment: 8 pages(Revtex), 5 figure

    On Fokker-Planck approximations of on-line learning processes

    Get PDF
    Contains fulltext : 100999.pdf (author's version ) (Open Access

    Bias/variance decompostion for likelihood-based estimators

    Get PDF
    Contains fulltext : 100975.pdf (author's version ) (Open Access

    Selecting weighting factors in logarithmic opinion pools

    Get PDF
    Contains fulltext : 100978.pdf (preprint version ) (Open Access

    Stochastics of on-line backpropagation

    Get PDF
    Contains fulltext : 101006.pdf (author's version ) (Open Access
    corecore