156,486 research outputs found
Inducing Interpretability in Knowledge Graph Embeddings
We study the problem of inducing interpretability in KG embeddings.
Specifically, we explore the Universal Schema (Riedel et al., 2013) and propose
a method to induce interpretability. There have been many vector space models
proposed for the problem, however, most of these methods don't address the
interpretability (semantics) of individual dimensions. In this work, we study
this problem and propose a method for inducing interpretability in KG
embeddings using entity co-occurrence statistics. The proposed method
significantly improves the interpretability, while maintaining comparable
performance in other KG tasks
Evaluating Sparse Interpretable Word Embeddings for Biomedical Domain
Word embeddings have found their way into a wide range of natural language
processing tasks including those in the biomedical domain. While these vector
representations successfully capture semantic and syntactic word relations,
hidden patterns and trends in the data, they fail to offer interpretability.
Interpretability is a key means to justification which is an integral part when
it comes to biomedical applications. We present an inclusive study on
interpretability of word embeddings in the medical domain, focusing on the role
of sparse methods. Qualitative and quantitative measurements and metrics for
interpretability of word vector representations are provided. For the
quantitative evaluation, we introduce an extensive categorized dataset that can
be used to quantify interpretability based on category theory. Intrinsic and
extrinsic evaluation of the studied methods are also presented. As for the
latter, we propose datasets which can be utilized for effective extrinsic
evaluation of word vectors in the biomedical domain. Based on our experiments,
it is seen that sparse word vectors show far more interpretability while
preserving the performance of their original vectors in downstream tasks
Interpretability via Model Extraction
The ability to interpret machine learning models has become increasingly
important now that machine learning is used to inform consequential decisions.
We propose an approach called model extraction for interpreting complex,
blackbox models. Our approach approximates the complex model using a much more
interpretable model; as long as the approximation quality is good, then
statistical properties of the complex model are reflected in the interpretable
model. We show how model extraction can be used to understand and debug random
forests and neural nets trained on several datasets from the UCI Machine
Learning Repository, as well as control policies learned for several classical
reinforcement learning problems.Comment: Presented as a poster at the 2017 Workshop on Fairness,
Accountability, and Transparency in Machine Learning (FAT/ML 2017
Machine Learning Interpretability: A Science rather than a tool
The term "interpretability" is oftenly used by machine learning researchers
each with their own intuitive understanding of it. There is no universal well
agreed upon definition of interpretability in machine learning. As any type of
science discipline is mainly driven by the set of formulated questions rather
than by different tools in that discipline, e.g. astrophysics is the discipline
that learns the composition of stars, not as the discipline that use the
spectroscopes. Similarly, we propose that machine learning interpretability
should be a discipline that answers specific questions related to
interpretability. These questions can be of statistical, causal and
counterfactual nature. Therefore, there is a need to look into the
interpretability problem of machine learning in the context of questions that
need to be addressed rather than different tools. We discuss about a
hypothetical interpretability framework driven by a question based scientific
approach rather than some specific machine learning model. Using a question
based notion of interpretability, we can step towards understanding the science
of machine learning rather than its engineering. This notion will also help us
understanding any specific problem more in depth rather than relying solely on
machine learning methods
Labelled tableaux for interpretability logics
In is paper we present a labelled tableau proof system that serves a wide
class of interpretability logics. The system is proved sound and complete for
any interpretability logic characterised by a frame condition given by a set of
universal strict first order Horn sentences. As such, the current paper adds to
a better proof-theoretical understanding of interpretability logics.Comment: Dedicated to Albert Visser on the occasion of his retirement. In:
Liber Amicorum Alberti, A Tribute to Albert Visser, Eds. Jan van Eijck,
Rosalie Iemhoff and Joost J. Joosten, p. 141-154, Tributes Series Vol. 30,
College Publications, London. ISBN 978-1-84890-204-6, 201
Computable functors and effective interpretability
Our main result is the equivalence of two notions of reducibility between
structures. One is a syntactical notion which is an effective version of
interpretability as in model theory, and the other one is a computational
notion which is a strengthening of the well-known Medvedev reducibility. We
extend our result to effective bi-interpretability and also to effective
reductions between classes of structures.Comment: 22 page
Learning Interpretable Features via Adversarially Robust Optimization
Neural networks are proven to be remarkably successful for classification and
diagnosis in medical applications. However, the ambiguity in the
decision-making process and the interpretability of the learned features is a
matter of concern. In this work, we propose a method for improving the feature
interpretability of neural network classifiers. Initially, we propose a
baseline convolutional neural network with state of the art performance in
terms of accuracy and weakly supervised localization. Subsequently, the loss is
modified to integrate robustness to adversarial examples into the training
process. In this work, feature interpretability is quantified via evaluating
the weakly supervised localization using the ground truth bounding boxes.
Interpretability is also visually assessed using class activation maps and
saliency maps. The method is applied to NIH ChestX-ray14, the largest publicly
available chest x-rays dataset. We demonstrate that the adversarially robust
optimization paradigm improves feature interpretability both quantitatively and
visually.Comment: MICCAI 2019 (Medical Image Computing and Computer Assisted
Interventions
Pseudosaturation and the Interpretability Orders
We streamline treatments of the interpretability orders
of Shelah, the key new notion being that of
pseudosaturation. Extending work of Malliaris and Shelah, we classify the
interpretability orders on the stable theories. As a further application, we
prove that for all countable theories , if is unsupersimple,
then if and only if . We thus deduce that simplicity is a dividing
line in , and that consistently,
characterizes maximality in ; previously these
results were only known for .Comment: 42 page
Beyond Explainability: Leveraging Interpretability for Improved Adversarial Learning
In this study, we propose the leveraging of interpretability for tasks beyond
purely the purpose of explainability. In particular, this study puts forward a
novel strategy for leveraging gradient-based interpretability in the realm of
adversarial examples, where we use insights gained to aid adversarial learning.
More specifically, we introduce the concept of spatially constrained one-pixel
adversarial perturbations, where we guide the learning of such adversarial
perturbations towards more susceptible areas identified via gradient-based
interpretability. Experimental results using different benchmark datasets show
that such a spatially constrained one-pixel adversarial perturbation strategy
can noticeably improve the speed of convergence as well as produce successful
attacks that were also visually difficult to perceive, thus illustrating an
effective use of interpretability methods for tasks outside of the purpose of
purely explainability.Comment: CVPR 2019 XAI Workshop accepte
How to improve the interpretability of kernel learning
In recent years, machine learning researchers have focused on methods to
construct flexible and interpretable prediction models. However, an
interpretability evaluation, a relationship between generalization performance
and an interpretability of the model and a method for improving the
interpretability have to be considered. In this paper, a quantitative index of
the interpretability is proposed and its rationality is proved, and equilibrium
problem between the interpretability and the generalization performance is
analyzed. Probability upper bound of the sum of the two performances is
analyzed. For traditional supervised kernel machine learning problem, a
universal learning framework is put forward to solve the equilibrium problem
between the two performances. The condition for global optimal solution based
on the framework is deduced. The learning framework is applied to the
least-squares support vector machine and is evaluated by some experiments.Comment: arXiv admin note: text overlap with arXiv:1811.0774
- …
