120 research outputs found
Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models
Shapley values underlie one of the most popular model-agnostic methods within
explainable artificial intelligence. These values are designed to attribute the
difference between a model's prediction and an average baseline to the
different features used as input to the model. Being based on solid
game-theoretic principles, Shapley values uniquely satisfy several desirable
properties, which is why they are increasingly used to explain the predictions
of possibly complex and highly non-linear machine learning models. Shapley
values are well calibrated to a user's intuition when features are independent,
but may lead to undesirable, counterintuitive explanations when the
independence assumption is violated.
In this paper, we propose a novel framework for computing Shapley values that
generalizes recent work that aims to circumvent the independence assumption. By
employing Pearl's do-calculus, we show how these 'causal' Shapley values can be
derived for general causal graphs without sacrificing any of their desirable
properties. Moreover, causal Shapley values enable us to separate the
contribution of direct and indirect effects. We provide a practical
implementation for computing causal Shapley values based on causal chain graphs
when only partial information is available and illustrate their utility on a
real-world example.Comment: Accepted at 34th Conference on Neural Information Processing Systems
(NeurIPS 2020
Learning Instantiation in First-Order Logic
Contains fulltext :
286055.pdf (Publisher’s version ) (Open Access)AITP 202
Dynamical and Stationary Properties of On-line Learning from Finite Training Sets
The dynamical and stationary properties of on-line learning from finite
training sets are analysed using the cavity method. For large input dimensions,
we derive equations for the macroscopic parameters, namely, the student-teacher
correlation, the student-student autocorrelation and the learning force
uctuation. This enables us to provide analytical solutions to Adaline learning
as a benchmark. Theoretical predictions of training errors in transient and
stationary states are obtained by a Monte Carlo sampling procedure.
Generalization and training errors are found to agree with simulations. The
physical origin of the critical learning rate is presented. Comparison with
batch learning is discussed throughout the paper.Comment: 30 pages, 4 figure
Investigation of topographical stability of the concave and convex Self-Organizing Map variant
We investigate, by a systematic numerical study, the parameter dependence of
the stability of the Kohonen Self-Organizing Map and the Zheng and Greenleaf
concave and convex learning with respect to different input distributions,
input and output dimensions
Effect of time-correlation of input patterns on the convergence of on-line learning
We studied the effects of time correlation of subsequent patterns on the
convergence of on-line learning by a feedforward neural network with
backpropagation algorithm. By using chaotic time series as sequences of
correlated patterns, we found that the unexpected scaling of converging time
with learning parameter emerges when time-correlated patterns accelerate
learning process.Comment: 8 pages(Revtex), 5 figure
On Fokker-Planck approximations of on-line learning processes
Contains fulltext :
100999.pdf (author's version ) (Open Access
Bias/variance decompostion for likelihood-based estimators
Contains fulltext :
100975.pdf (author's version ) (Open Access
Selecting weighting factors in logarithmic opinion pools
Contains fulltext :
100978.pdf (preprint version ) (Open Access
Stochastics of on-line backpropagation
Contains fulltext :
101006.pdf (author's version ) (Open Access
- …