1,080 research outputs found
Explaining Machine Learning Classifiers through Diverse Counterfactual Explanations
Post-hoc explanations of machine learning models are crucial for people to
understand and act on algorithmic predictions. An intriguing class of
explanations is through counterfactuals, hypothetical examples that show people
how to obtain a different prediction. We posit that effective counterfactual
explanations should satisfy two properties: feasibility of the counterfactual
actions given user context and constraints, and diversity among the
counterfactuals presented. To this end, we propose a framework for generating
and evaluating a diverse set of counterfactual explanations based on
determinantal point processes. To evaluate the actionability of
counterfactuals, we provide metrics that enable comparison of
counterfactual-based methods to other local explanation methods. We further
address necessary tradeoffs and point to causal implications in optimizing for
counterfactuals. Our experiments on four real-world datasets show that our
framework can generate a set of counterfactuals that are diverse and well
approximate local decision boundaries, outperforming prior approaches to
generating diverse counterfactuals. We provide an implementation of the
framework at https://github.com/microsoft/DiCE.Comment: 13 page
Algorithms to estimate Shapley value feature attributions
Feature attributions based on the Shapley value are popular for explaining
machine learning models; however, their estimation is complex from both a
theoretical and computational standpoint. We disentangle this complexity into
two factors: (1)~the approach to removing feature information, and (2)~the
tractable estimation strategy. These two factors provide a natural lens through
which we can better understand and compare 24 distinct algorithms. Based on the
various feature removal approaches, we describe the multiple types of Shapley
value feature attributions and methods to calculate each one. Then, based on
the tractable estimation strategies, we characterize two distinct families of
approaches: model-agnostic and model-specific approximations. For the
model-agnostic approximations, we benchmark a wide class of estimation
approaches and tie them to alternative yet equivalent characterizations of the
Shapley value. For the model-specific approximations, we clarify the
assumptions crucial to each method's tractability for linear, tree, and deep
models. Finally, we identify gaps in the literature and promising future
research directions
Novel Microdialysis Technique Reveals a Dramatic Shift in Metabolite Secretion during the Early Stages of the Interaction between the Ectomycorrhizal Fungus Pisolithus microcarpus and Its Host Eucalyptus grandis
The colonisation of tree roots by ectomycorrhizal (ECM) fungi is the result of numerous signalling exchanges between organisms, many of which occur before physical contact. However, information is lacking about these exchanges and the compounds that are secreted by each organism before contact. This is in part due to a lack of low disturbance sampling methods with sufficient temporal and spatial resolution to capture these exchanges. Using a novel in situ microdialysis approach, we sampled metabolites released from Eucalyptus grandis and Pisolithus microcarpus independently and during indirect contact over a 48-h time-course using UPLC-MS. A total of 560 and 1530 molecular features (MFs; ESI- and ESI+ respectively) were identified with significant differential abundance from control treatments. We observed that indirect contact between organisms altered the secretion of MFs to produce a distinct metabolomic profile compared to either organism independently. Many of these MFs were produced within the first hour of contact and included several phenylpropanoids, fatty acids and organic acids. These findings show that the secreted metabolome, particularly of the ECM fungus, can rapidly shift during the early stages of pre-symbiotic contact and highlight the importance of observing these early interactions in greater detail. We present microdialysis as a useful tool for examining plant-fungal signalling with high temporal resolution and with minimal experimental disturbance
On Human Predictions with Explanations and Predictions of Machine Learning Models: A Case Study on Deception Detection
Humans are the final decision makers in critical tasks that involve ethical
and legal concerns, ranging from recidivism prediction, to medical diagnosis,
to fighting against fake news. Although machine learning models can sometimes
achieve impressive performance in these tasks, these tasks are not amenable to
full automation. To realize the potential of machine learning for improving
human decisions, it is important to understand how assistance from machine
learning models affects human performance and human agency.
In this paper, we use deception detection as a testbed and investigate how we
can harness explanations and predictions of machine learning models to improve
human performance while retaining human agency. We propose a spectrum between
full human agency and full automation, and develop varying levels of machine
assistance along the spectrum that gradually increase the influence of machine
predictions. We find that without showing predicted labels, explanations alone
slightly improve human performance in the end task. In comparison, human
performance is greatly improved by showing predicted labels (>20% relative
improvement) and can be further improved by explicitly suggesting strong
machine performance. Interestingly, when predicted labels are shown,
explanations of machine predictions induce a similar level of accuracy as an
explicit statement of strong machine performance. Our results demonstrate a
tradeoff between human performance and human agency and show that explanations
of machine predictions can moderate this tradeoff.Comment: 17 pages, 19 figures, in Proceedings of ACM FAT* 2019, dataset & demo
available at https://deception.machineintheloop.co
Why Does My Model Fail? Contrastive Local Explanations for Retail Forecasting
In various business settings, there is an interest in using more complex
machine learning techniques for sales forecasting. It is difficult to convince
analysts, along with their superiors, to adopt these techniques since the
models are considered to be "black boxes," even if they perform better than
current models in use. We examine the impact of contrastive explanations about
large errors on users' attitudes towards a "black-box'" model. We propose an
algorithm, Monte Carlo Bounds for Reasonable Predictions. Given a large error,
MC-BRP determines (1) feature values that would result in a reasonable
prediction, and (2) general trends between each feature and the target, both
based on Monte Carlo simulations. We evaluate on a real dataset with real users
by conducting a user study with 75 participants to determine if explanations
generated by MC-BRP help users understand why a prediction results in a large
error, and if this promotes trust in an automatically-learned model. Our study
shows that users are able to answer objective questions about the model's
predictions with overall 81.1% accuracy when provided with these contrastive
explanations. We show that users who saw MC-BRP explanations understand why the
model makes large errors in predictions significantly more than users in the
control group. We also conduct an in-depth analysis on the difference in
attitudes between Practitioners and Researchers, and confirm that our results
hold when conditioning on the users' background.Comment: To appear in ACM FAT* 202
MT-Toolbox: improved amplicon sequencing using molecule tags
Abstract Background Short oligonucleotides can be used as markers to tag and track DNA sequences. For example, barcoding techniques (i.e. Multiplex Identifiers or Indexing) use short oligonucleotides to distinguish between reads from different DNA samples pooled for high-throughput sequencing. A similar technique called molecule tagging uses the same principles but is applied to individual DNA template molecules. Each template molecule is tagged with a unique oligonucleotide prior to polymerase chain reaction. The resulting amplicon sequences can be traced back to their original templates by their oligonucleotide tag. Consensus building from sequences sharing the same tag enables inference of original template molecules thereby reducing effects of sequencing error and polymerase chain reaction bias. Several independent groups have developed similar protocols for molecule tagging; however, user-friendly software for build consensus sequences from molecule tagged reads is not readily available or is highly specific for a particular protocol. Results MT-Toolbox recognizes oligonucleotide tags in amplicons and infers the correct template sequence. On a set of molecule tagged test reads, MT-Toolbox generates sequences having on average 0.00047 errors per base. MT-Toolbox includes a graphical user interface, command line interface, and options for speed and accuracy maximization. It can be run in serial on a standard personal computer or in parallel on a Load Sharing Facility based cluster system. An optional plugin provides features for common 16S metagenome profiling analysis such as chimera filtering, building operational taxonomic units, contaminant removal, and taxonomy assignments. Conclusions MT-Toolbox provides an accessible, user-friendly environment for analysis of molecule tagged reads thereby reducing technical errors and polymerase chain reaction bias. These improvements reduce noise and allow for greater precision in single amplicon sequencing experiments
- …