28 research outputs found
Neural Prototype Trees for Interpretable Fine-grained Image Recognition
Interpretable machine learning addresses the black-box nature of deep neural
networks. Visual prototypes have been suggested for intrinsically interpretable
image recognition, instead of generating post-hoc explanations that approximate
a trained model. However, a large number of prototypes can be overwhelming. To
reduce explanation size and improve interpretability, we propose the Neural
Prototype Tree (ProtoTree), a deep learning method that includes prototypes in
an interpretable decision tree to faithfully visualize the entire model. In
addition to global interpretability, a path in the tree explains a single
prediction. Each node in our binary tree contains a trainable prototypical
part. The presence or absence of this prototype in an image determines the
routing through a node. Decision making is therefore similar to human
reasoning: Does the bird have a red throat? And an elongated beak? Then it's a
hummingbird! We tune the accuracy-interpretability trade-off using ensembling
and pruning. We apply pruning without sacrificing accuracy, resulting in a
small tree with only 8 prototypes along a path to classify a bird from 200
species. An ensemble of 5 ProtoTrees achieves competitive accuracy on the
CUB-200-2011 and Stanford Cars data sets. Code is available at
https://github.com/M-Nauta/ProtoTreeComment: 11 pages, and 9 pages supplementar
This Looks Like That, Because ... Explaining Prototypes for Interpretable Image Recognition
Image recognition with prototypes is considered an interpretable alternative
for black box deep learning models. Classification depends on the extent to
which a test image "looks like" a prototype. However, perceptual similarity for
humans can be different from the similarity learned by the classification
model. Hence, only visualising prototypes can be insufficient for a user to
understand what a prototype exactly represents, and why the model considers a
prototype and an image to be similar. We address this ambiguity and argue that
prototypes should be explained. We improve interpretability by automatically
enhancing visual prototypes with textual quantitative information about visual
characteristics deemed important by the classification model. Specifically, our
method clarifies the meaning of a prototype by quantifying the influence of
colour hue, shape, texture, contrast and saturation and can generate both
global and local explanations. Because of the generality of our approach, it
can improve the interpretability of any similarity-based method for
prototypical image recognition. In our experiments, we apply our method to the
existing Prototypical Part Network (ProtoPNet). Our analysis confirms that the
global explanations are generalisable, and often correspond to the visually
perceptible properties of a prototype. Our explanations are especially relevant
for prototypes which might have been interpreted incorrectly otherwise. By
explaining such 'misleading' prototypes, we improve the interpretability and
simulatability of a prototype-based classification model. We also use our
method to check whether visually similar prototypes have similar explanations,
and are able to discover redundancy. Code is available at
https://github.com/M-Nauta/Explaining_Prototypes .Comment: 10 pages, 9 figure
Benchmarking eXplainable AI:A Survey on Available Toolkits and Open Challenges
The goal of Explainable AI (XAI) is to make the reasoning of a machine learning model accessible to humans, such that users of an AI system can evaluate and judge the underlying model. Due to the blackbox nature of XAI methods it is, however, hard to disentangle the contribution of a model and the explanation method to the final output. It might be unclear on whether an unexpected output is caused by the model or the explanation method. Explanation models, therefore, need to be evaluated in technical (e.g. fidelity to the model) and user-facing (correspondence to domain knowledge) terms. A recent survey has identified 29 different automated approaches to quantitatively evaluate explanations. In this work, we take an additional perspective and analyse which toolkits and data sets are available. We investigate which evaluation metrics are implemented in the toolkits and whether they produce the same results. We find that only a few aspects of explanation quality are currently covered, data sets are rare and evaluation results are not comparable across different toolkits. Our survey can serve as a guide for the XAI community for identifying future directions of research, and most notably, standardisation of evaluation.</p
Worst-Case Morphs using Wasserstein ALI and Improved MIPGAN
A morph is a combination of two separate facial images and contains identity
information of two different people. When used in an identity document, both
people can be authenticated by a biometric Face Recognition (FR) system. Morphs
can be generated using either a landmark-based approach or approaches based on
deep learning such as Generative Adversarial Networks (GAN). In a recent paper,
we introduced a \emph{worst-case} upper bound on how challenging morphing
attacks can be for an FR system. The closer morphs are to this upper bound, the
bigger the challenge they pose to FR. We introduced an approach with which it
was possible to generate morphs that approximate this upper bound for a known
FR system (white box), but not for unknown (black box) FR systems.
In this paper, we introduce a morph generation method that can approximate
worst-case morphs even when the FR system is not known. A key contribution is
that we include the goal of generating difficult morphs \emph{during} training.
Our method is based on Adversarially Learned Inference (ALI) and uses concepts
from Wasserstein GANs trained with Gradient Penalty, which were introduced to
stabilise the training of GANs. We include these concepts to achieve similar
improvement in training stability and call the resulting method Wasserstein ALI
(WALI). We finetune WALI using loss functions designed specifically to improve
the ability to manipulate identity information in facial images and show how it
can generate morphs that are more challenging for FR systems than landmark- or
GAN-based morphs. We also show how our findings can be used to improve MIPGAN,
an existing StyleGAN-based morph generator
Interpreting and Correcting Medical Image Classification with PIP-Net
Part-prototype models are explainable-by-design image classifiers, and a
promising alternative to black box AI. This paper explores the applicability
and potential of interpretable machine learning, in particular PIP-Net, for
automated diagnosis support on real-world medical imaging data. PIP-Net learns
human-understandable prototypical image parts and we evaluate its accuracy and
interpretability for fracture detection and skin cancer diagnosis. We find that
PIP-Net's decision making process is in line with medical classification
standards, while only provided with image-level class labels. Because of
PIP-Net's unsupervised pretraining of prototypes, data quality problems such as
undesired text in an X-ray or labelling errors can be easily identified.
Additionally, we are the first to show that humans can manually correct the
reasoning of PIP-Net by directly disabling undesired prototypes. We conclude
that part-prototype models are promising for medical applications due to their
interpretability and potential for advanced model debugging