20,908 research outputs found
Techniques for visualizing LSTMs applied to electrocardiograms
This paper explores four different visualization techniques for long
short-term memory (LSTM) networks applied to continuous-valued time series. On
the datasets analysed, we find that the best visualization technique is to
learn an input deletion mask that optimally reduces the true class score. With
a specific focus on single-lead electrocardiograms from the MIT-BIH arrhythmia
dataset, we show that salient input features for the LSTM classifier align well
with medical theory.Comment: presented at 2018 ICML Workshop on Human Interpretability in Machine
Learning (WHI 2018), Stockholm, Swede
Visualizing and Understanding Atari Agents
While deep reinforcement learning (deep RL) agents are effective at
maximizing rewards, it is often unclear what strategies they use to do so. In
this paper, we take a step toward explaining deep RL agents through a case
study using Atari 2600 environments. In particular, we focus on using saliency
maps to understand how an agent learns and executes a policy. We introduce a
method for generating useful saliency maps and use it to show 1) what strong
agents attend to, 2) whether agents are making decisions for the right or wrong
reasons, and 3) how agents evolve during learning. We also test our method on
non-expert human subjects and find that it improves their ability to reason
about these agents. Overall, our results show that saliency information can
provide significant insight into an RL agent's decisions and learning behavior.Comment: ICML 2018 conference paper. Code:
https://github.com/greydanus/visualize_atari Blog:
https://greydanus.github.io/2017/11/01/visualize-atari
LISA: Explaining Recurrent Neural Network Judgments via Layer-wIse Semantic Accumulation and Example to Pattern Transformation
Recurrent neural networks (RNNs) are temporal networks and cumulative in
nature that have shown promising results in various natural language processing
tasks. Despite their success, it still remains a challenge to understand their
hidden behavior. In this work, we analyze and interpret the cumulative nature
of RNN via a proposed technique named as Layer-wIse-Semantic-Accumulation
(LISA) for explaining decisions and detecting the most likely (i.e., saliency)
patterns that the network relies on while decision making. We demonstrate (1)
LISA: "How an RNN accumulates or builds semantics during its sequential
processing for a given text example and expected response" (2) Example2pattern:
"How the saliency patterns look like for each category in the data according to
the network in decision making". We analyse the sensitiveness of RNNs about
different inputs to check the increase or decrease in prediction scores and
further extract the saliency patterns learned by the network. We employ two
relation classification datasets: SemEval 10 Task 8 and TAC KBP Slot Filling to
explain RNN predictions via the LISA and example2pattern.Comment: 2018 Conference on Empirical Methods in Natural Language Processing
(EMNLP2018) workshop on Analyzing and Interpreting Neural Networks for NLP
(BlackBoxNLP
Visual Analytics in Deep Learning: An Interrogative Survey for the Next Frontiers
Deep learning has recently seen rapid development and received significant
attention due to its state-of-the-art performance on previously-thought hard
problems. However, because of the internal complexity and nonlinear structure
of deep neural networks, the underlying decision making processes for why these
models are achieving such performance are challenging and sometimes mystifying
to interpret. As deep learning spreads across domains, it is of paramount
importance that we equip users of deep learning with tools for understanding
when a model works correctly, when it fails, and ultimately how to improve its
performance. Standardized toolkits for building neural networks have helped
democratize deep learning; visual analytics systems have now been developed to
support model explanation, interpretation, debugging, and improvement. We
present a survey of the role of visual analytics in deep learning research,
which highlights its short yet impactful history and thoroughly summarizes the
state-of-the-art using a human-centered interrogative framework, focusing on
the Five W's and How (Why, Who, What, How, When, and Where). We conclude by
highlighting research directions and open research problems. This survey helps
researchers and practitioners in both visual analytics and deep learning to
quickly learn key aspects of this young and rapidly growing body of research,
whose impact spans a diverse range of domains.Comment: Under review for IEEE Transactions on Visualization and Computer
Graphics (TVCG
Attend and Predict: Understanding Gene Regulation by Selective Attention on Chromatin
The past decade has seen a revolution in genomic technologies that enable a
flood of genome-wide profiling of chromatin marks. Recent literature tried to
understand gene regulation by predicting gene expression from large-scale
chromatin measurements. Two fundamental challenges exist for such learning
tasks: (1) genome-wide chromatin signals are spatially structured,
high-dimensional and highly modular; and (2) the core aim is to understand what
are the relevant factors and how they work together? Previous studies either
failed to model complex dependencies among input signals or relied on separate
feature analysis to explain the decisions. This paper presents an
attention-based deep learning approach; we call AttentiveChrome, that uses a
unified architecture to model and to interpret dependencies among chromatin
factors for controlling gene regulation. AttentiveChrome uses a hierarchy of
multiple Long short-term memory (LSTM) modules to encode the input signals and
to model how various chromatin marks cooperate automatically. AttentiveChrome
trains two levels of attention jointly with the target prediction, enabling it
to attend differentially to relevant marks and to locate important positions
per mark. We evaluate the model across 56 different cell types (tasks) in
human. Not only is the proposed architecture more accurate, but its attention
scores also provide a better interpretation than state-of-the-art feature
visualization methods such as saliency map.
Code and data are shared at www.deepchrome.orgComment: 12 pages; At NIPS 201
Visual Reasoning of Feature Attribution with Deep Recurrent Neural Networks
Deep Recurrent Neural Network (RNN) has gained popularity in many sequence
classification tasks. Beyond predicting a correct class for each data instance,
data scientists also want to understand what differentiating factors in the
data have contributed to the classification during the learning process. We
present a visual analytics approach to facilitate this task by revealing the
RNN attention for all data instances, their temporal positions in the
sequences, and the attribution of variables at each value level. We demonstrate
with real-world datasets that our approach can help data scientists to
understand such dynamics in deep RNNs from the training results, hence guiding
their modeling process
Understanding Recurrent Neural State Using Memory Signatures
We demonstrate a network visualization technique to analyze the recurrent
state inside the LSTMs/GRUs used commonly in language and acoustic models.
Interpreting intermediate state and network activations inside end-to-end
models remains an open challenge. Our method allows users to understand exactly
how much and what history is encoded inside recurrent state in grapheme
sequence models. Our procedure trains multiple decoders that predict prior
input history. Compiling results from these decoders, a user can obtain a
signature of the recurrent kernel that characterizes its memory behavior. We
demonstrate this method's usefulness in revealing information divergence in the
bases of recurrent factorized kernels, visualizing the character-level
differences between the memory of n-gram and recurrent language models, and
extracting knowledge of history encoded in the layers of grapheme-based
end-to-end ASR networks.Comment: Accepted to 2018 IEEE International Conference on Acoustics, Speech
and Signal Processin
Recent Advances in Deep Learning: An Overview
Deep Learning is one of the newest trends in Machine Learning and Artificial
Intelligence research. It is also one of the most popular scientific research
trends now-a-days. Deep learning methods have brought revolutionary advances in
computer vision and machine learning. Every now and then, new and new deep
learning techniques are being born, outperforming state-of-the-art machine
learning and even existing deep learning techniques. In recent years, the world
has seen many major breakthroughs in this field. Since deep learning is
evolving at a huge speed, its kind of hard to keep track of the regular
advances especially for new researchers. In this paper, we are going to briefly
discuss about recent advances in Deep Learning for past few years.Comment: 31 pages including bibliograph
RetainVis: Visual Analytics with Interpretable and Interactive Recurrent Neural Networks on Electronic Medical Records
We have recently seen many successful applications of recurrent neural
networks (RNNs) on electronic medical records (EMRs), which contain histories
of patients' diagnoses, medications, and other various events, in order to
predict the current and future states of patients. Despite the strong
performance of RNNs, it is often challenging for users to understand why the
model makes a particular prediction. Such black-box nature of RNNs can impede
its wide adoption in clinical practice. Furthermore, we have no established
methods to interactively leverage users' domain expertise and prior knowledge
as inputs for steering the model. Therefore, our design study aims to provide a
visual analytics solution to increase interpretability and interactivity of
RNNs via a joint effort of medical experts, artificial intelligence scientists,
and visual analytics researchers. Following the iterative design process
between the experts, we design, implement, and evaluate a visual analytics tool
called RetainVis, which couples a newly improved, interpretable and interactive
RNN-based model called RetainEX and visualizations for users' exploration of
EMR data in the context of prediction tasks. Our study shows the effective use
of RetainVis for gaining insights into how individual medical codes contribute
to making risk predictions, using EMRs of patients with heart failure and
cataract symptoms. Our study also demonstrates how we made substantial changes
to the state-of-the-art RNN model called RETAIN in order to make use of
temporal information and increase interactivity. This study will provide a
useful guideline for researchers that aim to design an interpretable and
interactive visual analytics tool for RNNs.Comment: Accepted at IEEE VIS 2018. To appear in IEEE Transactions on
Visualization and Computer Graphics in January 201
Increasing the Interpretability of Recurrent Neural Networks Using Hidden Markov Models
As deep neural networks continue to revolutionize various application
domains, there is increasing interest in making these powerful models more
understandable and interpretable, and narrowing down the causes of good and bad
predictions. We focus on recurrent neural networks (RNNs), state of the art
models in speech recognition and translation. Our approach to increasing
interpretability is by combining an RNN with a hidden Markov model (HMM), a
simpler and more transparent model. We explore various combinations of RNNs and
HMMs: an HMM trained on LSTM states; a hybrid model where an HMM is trained
first, then a small LSTM is given HMM state distributions and trained to fill
in gaps in the HMM's performance; and a jointly trained hybrid model. We find
that the LSTM and HMM learn complementary information about the features in the
text.Comment: presented at 2016 ICML Workshop on Human Interpretability in Machine
Learning (WHI 2016), New York, N
- …