127 research outputs found
Iterative Partial Fulfillment of Counterfactual Explanations: Benefits and Risks
Counterfactual (CF) explanations, also known as contrastive explanations and
algorithmic recourses, are popular for explaining machine learning models in
high-stakes domains. For a subject that receives a negative model prediction
(e.g., mortgage application denial), the CF explanations are similar instances
but with positive predictions, which informs the subject of ways to improve.
While their various properties have been studied, such as validity and
stability, we contribute a novel one: their behaviors under iterative partial
fulfillment (IPF). Specifically, upon receiving a CF explanation, the subject
may only partially fulfill it before requesting a new prediction with a new
explanation, and repeat until the prediction is positive. Such partial
fulfillment could be due to the subject's limited capability (e.g., can only
pay down two out of four credit card accounts at this moment) or an attempt to
take the chance (e.g., betting that a monthly salary increase of \$800 is
enough even though \$1,000 is recommended). Does such iterative partial
fulfillment increase or decrease the total cost of improvement incurred by the
subject? We mathematically formalize IPF and demonstrate, both theoretically
and empirically, that different CF algorithms exhibit vastly different
behaviors under IPF. We discuss implications of our observations, advocate for
this factor to be carefully considered in the development and study of CF
algorithms, and give several directions for future work.Comment: AIES 202
Predicting ConceptNet Path Quality Using Crowdsourced Assessments of Naturalness
In many applications, it is important to characterize the way in which two
concepts are semantically related. Knowledge graphs such as ConceptNet provide
a rich source of information for such characterizations by encoding relations
between concepts as edges in a graph. When two concepts are not directly
connected by an edge, their relationship can still be described in terms of the
paths that connect them. Unfortunately, many of these paths are uninformative
and noisy, which means that the success of applications that use such path
features crucially relies on their ability to select high-quality paths. In
existing applications, this path selection process is based on relatively
simple heuristics. In this paper we instead propose to learn to predict path
quality from crowdsourced human assessments. Since we are interested in a
generic task-independent notion of quality, we simply ask human participants to
rank paths according to their subjective assessment of the paths' naturalness,
without attempting to define naturalness or steering the participants towards
particular indicators of quality. We show that a neural network model trained
on these assessments is able to predict human judgments on unseen paths with
near optimal performance. Most notably, we find that the resulting path
selection method is substantially better than the current heuristic approaches
at identifying meaningful paths.Comment: In Proceedings of the Web Conference (WWW) 201
Improving Generalization in Language Model-Based Text-to-SQL Semantic Parsing: Two Simple Semantic Boundary-Based Techniques
Compositional and domain generalization present significant challenges in
semantic parsing, even for state-of-the-art semantic parsers based on
pre-trained language models (LMs). In this study, we empirically investigate
improving an LM's generalization in semantic parsing with two simple
techniques: at the token level, we introduce a token preprocessing method to
preserve the semantic boundaries of tokens produced by LM tokenizers; at the
sequence level, we propose to use special tokens to mark the boundaries of
components aligned between input and output. Our experimental results on two
text-to-SQL semantic parsing datasets show that our token preprocessing,
although simple, can substantially improve the LM performance on both types of
generalization, and our component boundary marking method is particularly
helpful for compositional generalization.Comment: 9 pages, to be published in ACL202
Bayes-TrEx: a Bayesian Sampling Approach to Model Transparency by Example
Post-hoc explanation methods are gaining popularity for interpreting,
understanding, and debugging neural networks. Most analyses using such methods
explain decisions in response to inputs drawn from the test set. However, the
test set may have few examples that trigger some model behaviors, such as
high-confidence failures or ambiguous classifications. To address these
challenges, we introduce a flexible model inspection framework: Bayes-TrEx.
Given a data distribution, Bayes-TrEx finds in-distribution examples with a
specified prediction confidence. We demonstrate several use cases of
Bayes-TrEx, including revealing highly confident (mis)classifications,
visualizing class boundaries via ambiguous examples, understanding novel-class
extrapolation behavior, and exposing neural network overconfidence. We use
Bayes-TrEx to study classifiers trained on CLEVR, MNIST, and Fashion-MNIST, and
we show that this framework enables more flexible holistic model analysis than
just inspecting the test set. Code is available at
https://github.com/serenabooth/Bayes-TrEx.Comment: Accepted at AAAI 202
- …