474 research outputs found
SCOUT: Self-aware Discriminant Counterfactual Explanations
The problem of counterfactual visual explanations is considered. A new family
of discriminant explanations is introduced. These produce heatmaps that
attribute high scores to image regions informative of a classifier prediction
but not of a counter class. They connect attributive explanations, which are
based on a single heat map, to counterfactual explanations, which account for
both predicted class and counter class. The latter are shown to be computable
by combination of two discriminant explanations, with reversed class pairs. It
is argued that self-awareness, namely the ability to produce classification
confidence scores, is important for the computation of discriminant
explanations, which seek to identify regions where it is easy to discriminate
between prediction and counter class. This suggests the computation of
discriminant explanations by the combination of three attribution maps. The
resulting counterfactual explanations are optimization free and thus much
faster than previous methods. To address the difficulty of their evaluation, a
proxy task and set of quantitative metrics are also proposed. Experiments under
this protocol show that the proposed counterfactual explanations outperform the
state of the art while achieving much higher speeds, for popular networks. In a
human-learning machine teaching experiment, they are also shown to improve mean
student accuracy from chance level to 95\%.Comment: Accepted to CVPR202
One Explanation Does Not Fit All The Promise of Interactive Explanations for Machine Learning Transparency
The need for transparency of predictive systems based on Machine Learning
algorithms arises as a consequence of their ever-increasing proliferation in
the industry. Whenever black-box algorithmic predictions influence human
affairs, the inner workings of these algorithms should be scrutinised and their
decisions explained to the relevant stakeholders, including the system
engineers, the system's operators and the individuals whose case is being
decided. While a variety of interpretability and explainability methods is
available, none of them is a panacea that can satisfy all diverse expectations
and competing objectives that might be required by the parties involved. We
address this challenge in this paper by discussing the promises of Interactive
Machine Learning for improved transparency of black-box systems using the
example of contrastive explanations -- a state-of-the-art approach to
Interpretable Machine Learning.
Specifically, we show how to personalise counterfactual explanations by
interactively adjusting their conditional statements and extract additional
explanations by asking follow-up "What if?" questions. Our experience in
building, deploying and presenting this type of system allowed us to list
desired properties as well as potential limitations, which can be used to guide
the development of interactive explainers. While customising the medium of
interaction, i.e., the user interface comprising of various communication
channels, may give an impression of personalisation, we argue that adjusting
the explanation itself and its content is more important. To this end,
properties such as breadth, scope, context, purpose and target of the
explanation have to be considered, in addition to explicitly informing the
explainee about its limitations and caveats...Comment: Published in the Kunstliche Intelligenz journal, special issue on
Challenges in Interactive Machine Learnin
- …