3,215 research outputs found
An Operation Sequence Model for Explainable Neural Machine Translation
We propose to achieve explainable neural machine translation (NMT) by
changing the output representation to explain itself. We present a novel
approach to NMT which generates the target sentence by monotonically walking
through the source sentence. Word reordering is modeled by operations which
allow setting markers in the target sentence and move a target-side write head
between those markers. In contrast to many modern neural models, our system
emits explicit word alignment information which is often crucial to practical
machine translation as it improves explainability. Our technique can outperform
a plain text system in terms of BLEU score under the recent Transformer
architecture on Japanese-English and Portuguese-English, and is within 0.5 BLEU
difference on Spanish-English
Analyzing and Interpreting Neural Networks for NLP: A Report on the First BlackboxNLP Workshop
The EMNLP 2018 workshop BlackboxNLP was dedicated to resources and techniques
specifically developed for analyzing and understanding the inner-workings and
representations acquired by neural models of language. Approaches included:
systematic manipulation of input to neural networks and investigating the
impact on their performance, testing whether interpretable knowledge can be
decoded from intermediate representations acquired by neural networks,
proposing modifications to neural network architectures to make their knowledge
state or generated output more explainable, and examining the performance of
networks on simplified or formal languages. Here we review a number of
representative studies in each category
INTERACTION: A Generative XAI Framework for Natural Language Inference Explanations
XAI with natural language processing aims to produce human-readable
explanations as evidence for AI decision-making, which addresses explainability
and transparency. However, from an HCI perspective, the current approaches only
focus on delivering a single explanation, which fails to account for the
diversity of human thoughts and experiences in language. This paper thus
addresses this gap, by proposing a generative XAI framework, INTERACTION
(explaIn aNd predicT thEn queRy with contextuAl CondiTional varIational
autO-eNcoder). Our novel framework presents explanation in two steps: (step
one) Explanation and Label Prediction; and (step two) Diverse Evidence
Generation. We conduct intensive experiments with the Transformer architecture
on a benchmark dataset, e-SNLI. Our method achieves competitive or better
performance against state-of-the-art baseline models on explanation generation
(up to 4.7% gain in BLEU) and prediction (up to 4.4% gain in accuracy) in step
one; it can also generate multiple diverse explanations in step two
Modeling the Field Value Variations and Field Interactions Simultaneously for Fraud Detection
With the explosive growth of e-commerce, online transaction fraud has become
one of the biggest challenges for e-commerce platforms. The historical
behaviors of users provide rich information for digging into the users' fraud
risk. While considerable efforts have been made in this direction, a
long-standing challenge is how to effectively exploit internal user information
and provide explainable prediction results. In fact, the value variations of
same field from different events and the interactions of different fields
inside one event have proven to be strong indicators for fraudulent behaviors.
In this paper, we propose the Dual Importance-aware Factorization Machines
(DIFM), which exploits the internal field information among users' behavior
sequence from dual perspectives, i.e., field value variations and field
interactions simultaneously for fraud detection. The proposed model is deployed
in the risk management system of one of the world's largest e-commerce
platforms, which utilize it to provide real-time transaction fraud detection.
Experimental results on real industrial data from different regions in the
platform clearly demonstrate that our model achieves significant improvements
compared with various state-of-the-art baseline models. Moreover, the DIFM
could also give an insight into the explanation of the prediction results from
dual perspectives.Comment: 11 pages, 4 figure
Rationalization for Explainable NLP: A Survey
Recent advances in deep learning have improved the performance of many
Natural Language Processing (NLP) tasks such as translation,
question-answering, and text classification. However, this improvement comes at
the expense of model explainability. Black-box models make it difficult to
understand the internals of a system and the process it takes to arrive at an
output. Numerical (LIME, Shapley) and visualization (saliency heatmap)
explainability techniques are helpful; however, they are insufficient because
they require specialized knowledge. These factors led rationalization to emerge
as a more accessible explainable technique in NLP. Rationalization justifies a
model's output by providing a natural language explanation (rationale). Recent
improvements in natural language generation have made rationalization an
attractive technique because it is intuitive, human-comprehensible, and
accessible to non-technical users. Since rationalization is a relatively new
field, it is disorganized. As the first survey, rationalization literature in
NLP from 2007-2022 is analyzed. This survey presents available methods,
explainable evaluations, code, and datasets used across various NLP tasks that
use rationalization. Further, a new subfield in Explainable AI (XAI), namely,
Rational AI (RAI), is introduced to advance the current state of
rationalization. A discussion on observed insights, challenges, and future
directions is provided to point to promising research opportunities
The blessings of explainable AI in operations & maintenance of wind turbines
Wind turbines play an integral role in generating clean energy, but regularly suffer from operational inconsistencies and failures leading to unexpected downtimes and significant Operations & Maintenance (O&M) costs. Condition-Based Monitoring (CBM) has been utilised in the past to monitor operational inconsistencies in turbines by applying signal processing techniques to vibration data. The last decade has witnessed growing interest in leveraging Supervisory Control & Acquisition (SCADA) data from turbine sensors towards CBM. Machine Learning (ML) techniques have been utilised to predict incipient faults in turbines and forecast vital operational parameters with high accuracy by leveraging SCADA data and alarm logs. More recently, Deep Learning (DL) methods have outperformed conventional ML techniques, particularly for anomaly prediction. Despite demonstrating immense promise in transitioning to Artificial Intelligence (AI), such models are generally black-boxes that cannot provide rationales behind their predictions, hampering the ability of turbine operators to rely on automated decision making. We aim to help combat this challenge by providing a novel perspective on Explainable AI (XAI) for trustworthy decision support.This thesis revolves around three key strands of XAI – DL, Natural Language Generation (NLG) and Knowledge Graphs (KGs), which are investigated by utilising data from an operational turbine. We leverage DL and NLG to predict incipient faults and alarm events in the turbine in natural language as well as generate human-intelligible O&M strategies to assist engineers in fixing/averting the faults. We also propose specialised DL models which can predict causal relationships in SCADA features as well as quantify the importance of vital parameters leading to failures. The thesis finally culminates with an interactive Question- Answering (QA) system for automated reasoning that leverages multimodal domain-specific information from a KG, facilitating engineers to retrieve O&M strategies with natural language questions. By helping make turbines more reliable, we envisage wider adoption of wind energy sources towards tackling climate change
xDBTagger: Explainable Natural Language Interface to Databases Using Keyword Mappings and Schema Graph
Translating natural language queries (NLQ) into structured query language
(SQL) in interfaces to relational databases is a challenging task that has been
widely studied by researchers from both the database and natural language
processing communities. Numerous works have been proposed to attack the natural
language interfaces to databases (NLIDB) problem either as a conventional
pipeline-based or an end-to-end deep-learning-based solution. Nevertheless,
regardless of the approach preferred, such solutions exhibit black-box nature,
which makes it difficult for potential users targeted by these systems to
comprehend the decisions made to produce the translated SQL. To this end, we
propose xDBTagger, an explainable hybrid translation pipeline that explains the
decisions made along the way to the user both textually and visually. We also
evaluate xDBTagger quantitatively in three real-world relational databases. The
evaluation results indicate that in addition to being fully interpretable,
xDBTagger is effective in terms of accuracy and translates the queries more
efficiently compared to other state-of-the-art pipeline-based systems up to
10000 times.Comment: 20 pages, 6 figures. This work is the extended version of
arXiv:2101.04226 that appeared in PVLDB'2
- …