40 research outputs found
Explainable AI and Interpretable Computer Vision:From Oversight to Insight
The increasing availability of big data and computational power has facilitated unprecedented progress in Artificial Intelligence (AI) and Machine Learning (ML). However, complex model architectures have resulted in high-performing yet uninterpretable âblack boxesâ. This prevents users from verifying that the reasoning process aligns with expectations and intentions. This thesis posits that the sole focus on predictive performance is an unsustainable trajectory, since a model can make right predictions for the wrong reasons. The research field of Explainable AI (XAI) addresses the black-box nature of AI by generating explanations that present (aspects of) a model's behaviour in human-understandable terms. This thesis supports the transition from oversight to insight, and shows that explainability can give users more insight into every part of the machine learning pipeline: from the training data to the prediction model and the resulting explanations. When relying on explanations for judging a model's reasoning process, it is important that the explanations are truthful, relevant and understandable. Part I of this thesis reflects upon explanation quality and identifies 12 desirable properties, including compactness, completeness and correctness. Additionally, it provides an extensive collection of quantitative XAI evaluation methods, and analyses their availabilities in open-source toolkits. As alternative to common post-model explainability that reverse-engineers an already trained prediction model, Part II of this thesis presents in-model explainability for interpretable computer vision. These image classifiers learn prototypical parts, which are used in an interpretable decision tree or scoring sheet. The models are explainable by design since their reasoning depends on the extent to which an image patch âlooks likeâ a learned part-prototype. Part III of this thesis shows that ML can also explain characteristics of a dataset. Because of a model's ability to analyse large amounts of data in little time, extracting hidden patterns can contribute to the validation and potential discovery of domain knowledge, and allows to detect sources of bias and shortcuts early on. Concluding, neither the prediction model nor the data nor the explanation method should be handled as a black box. The way forward? AI with a human touch: developing powerful models that learn interpretable features, and using these meaningful features in a decision process that users can understand, validate and adapt. This in-model explainability, such as the part-prototype models from Part II, opens up the opportunity to âre-educateâ models with our desired norms, values and reasoning. Enabling human decision-makers to detect and correct undesired model behaviour will contribute towards an effective but also reliable and responsible usage of AI
Neural Prototype Trees for Interpretable Fine-grained Image Recognition
Interpretable machine learning addresses the black-box nature of deep neural
networks. Visual prototypes have been suggested for intrinsically interpretable
image recognition, instead of generating post-hoc explanations that approximate
a trained model. However, a large number of prototypes can be overwhelming. To
reduce explanation size and improve interpretability, we propose the Neural
Prototype Tree (ProtoTree), a deep learning method that includes prototypes in
an interpretable decision tree to faithfully visualize the entire model. In
addition to global interpretability, a path in the tree explains a single
prediction. Each node in our binary tree contains a trainable prototypical
part. The presence or absence of this prototype in an image determines the
routing through a node. Decision making is therefore similar to human
reasoning: Does the bird have a red throat? And an elongated beak? Then it's a
hummingbird! We tune the accuracy-interpretability trade-off using ensembling
and pruning. We apply pruning without sacrificing accuracy, resulting in a
small tree with only 8 prototypes along a path to classify a bird from 200
species. An ensemble of 5 ProtoTrees achieves competitive accuracy on the
CUB-200-2011 and Stanford Cars data sets. Code is available at
https://github.com/M-Nauta/ProtoTreeComment: 11 pages, and 9 pages supplementar
Feature Attribution Explanations for Spiking Neural Networks
Third-generation artificial neural networks, Spiking Neural Networks (SNNs), can be efficiently implemented on hardware. Their implementation on neuromorphic chips opens a broad range of applications, such as machine learning-based autonomous control and intelligent biomedical devices. In critical applications, however, insight into the reasoning of SNNs is important, thus SNNs need to be equipped with the ability to explain how decisions are reached. We present \textit{Temporal Spike Attribution} (TSA), a local explanation method for SNNs. To compute the explanation, we aggregate all information available in model-internal variables: spike times and model weights. We evaluate TSA on artificial and real-world time series data and measure explanation quality w.r.t. multiple quantitative criteria. We find that TSA correctly identifies a small subset of input features relevant to the decision (i.e., is output-complete and compact) and generates similar explanations for similar inputs (i.e., is continuous). Further, our experiments show that incorporating the notion of \emph{absent} spikes improves explanation quality. Our work can serve as a starting point for explainable SNNs, with future implementations on hardware yielding not only predictions but also explanations in a broad range of application scenarios. Source code is available at https://github.com/ElisaNguyen/tsa-explanations
This Looks Like That, Because ... Explaining Prototypes for Interpretable Image Recognition
Image recognition with prototypes is considered an interpretable alternative
for black box deep learning models. Classification depends on the extent to
which a test image "looks like" a prototype. However, perceptual similarity for
humans can be different from the similarity learned by the classification
model. Hence, only visualising prototypes can be insufficient for a user to
understand what a prototype exactly represents, and why the model considers a
prototype and an image to be similar. We address this ambiguity and argue that
prototypes should be explained. We improve interpretability by automatically
enhancing visual prototypes with textual quantitative information about visual
characteristics deemed important by the classification model. Specifically, our
method clarifies the meaning of a prototype by quantifying the influence of
colour hue, shape, texture, contrast and saturation and can generate both
global and local explanations. Because of the generality of our approach, it
can improve the interpretability of any similarity-based method for
prototypical image recognition. In our experiments, we apply our method to the
existing Prototypical Part Network (ProtoPNet). Our analysis confirms that the
global explanations are generalisable, and often correspond to the visually
perceptible properties of a prototype. Our explanations are especially relevant
for prototypes which might have been interpreted incorrectly otherwise. By
explaining such 'misleading' prototypes, we improve the interpretability and
simulatability of a prototype-based classification model. We also use our
method to check whether visually similar prototypes have similar explanations,
and are able to discover redundancy. Code is available at
https://github.com/M-Nauta/Explaining_Prototypes .Comment: 10 pages, 9 figure
Worst-Case Morphs using Wasserstein ALI and Improved MIPGAN
A morph is a combination of two separate facial images and contains identity information of two different people. When used in an identity document, both people can be authenticated by a biometric Face Recognition (FR) system. Morphs can be generated using either a landmark-based approach or approaches based on deep learning such as Generative Adversarial Networks (GAN). In a recent paper, we introduced a \emph{worst-case} upper bound on how challenging morphing attacks can be for an FR system. The closer morphs are to this upper bound, the bigger the challenge they pose to FR. We introduced an approach with which it was possible to generate morphs that approximate this upper bound for a known FR system (white box), but not for unknown (black box) FR systems. In this paper, we introduce a morph generation method that can approximate worst-case morphs even when the FR system is not known. A key contribution is that we include the goal of generating difficult morphs \emph{during} training. Our method is based on Adversarially Learned Inference (ALI) and uses concepts from Wasserstein GANs trained with Gradient Penalty, which were introduced to stabilise the training of GANs. We include these concepts to achieve similar improvement in training stability and call the resulting method Wasserstein ALI (WALI). We finetune WALI using loss functions designed specifically to improve the ability to manipulate identity information in facial images and show how it can generate morphs that are more challenging for FR systems than landmark- or GAN-based morphs. We also show how our findings can be used to improve MIPGAN, an existing StyleGAN-based morph generator
Worst-Case Morphs using Wasserstein ALI and Improved MIPGAN
A morph is a combination of two separate facial images and contains identity
information of two different people. When used in an identity document, both
people can be authenticated by a biometric Face Recognition (FR) system. Morphs
can be generated using either a landmark-based approach or approaches based on
deep learning such as Generative Adversarial Networks (GAN). In a recent paper,
we introduced a \emph{worst-case} upper bound on how challenging morphing
attacks can be for an FR system. The closer morphs are to this upper bound, the
bigger the challenge they pose to FR. We introduced an approach with which it
was possible to generate morphs that approximate this upper bound for a known
FR system (white box), but not for unknown (black box) FR systems.
In this paper, we introduce a morph generation method that can approximate
worst-case morphs even when the FR system is not known. A key contribution is
that we include the goal of generating difficult morphs \emph{during} training.
Our method is based on Adversarially Learned Inference (ALI) and uses concepts
from Wasserstein GANs trained with Gradient Penalty, which were introduced to
stabilise the training of GANs. We include these concepts to achieve similar
improvement in training stability and call the resulting method Wasserstein ALI
(WALI). We finetune WALI using loss functions designed specifically to improve
the ability to manipulate identity information in facial images and show how it
can generate morphs that are more challenging for FR systems than landmark- or
GAN-based morphs. We also show how our findings can be used to improve MIPGAN,
an existing StyleGAN-based morph generator