104 research outputs found
Interpretability and Explainability: A Machine Learning Zoo Mini-tour
In this review, we examine the problem of designing interpretable and
explainable machine learning models. Interpretability and explainability lie at
the core of many machine learning and statistical applications in medicine,
economics, law, and natural sciences. Although interpretability and
explainability have escaped a clear universal definition, many techniques
motivated by these properties have been developed over the recent 30 years with
the focus currently shifting towards deep learning methods. In this review, we
emphasise the divide between interpretability and explainability and illustrate
these two different research directions with concrete examples of the
state-of-the-art. The review is intended for a general machine learning
audience with interest in exploring the problems of interpretation and
explanation beyond logistic regression or random forest variable importance.
This work is not an exhaustive literature survey, but rather a primer focusing
selectively on certain lines of research which the authors found interesting or
informative
A Modular Task-oriented Dialogue System Using a Neural Mixture-of-Experts
End-to-end Task-oriented Dialogue Systems (TDSs) have attracted a lot of
attention for their superiority (e.g., in terms of global optimization) over
pipeline modularized TDSs. Previous studies on end-to-end TDSs use a
single-module model to generate responses for complex dialogue contexts.
However, no model consistently outperforms the others in all cases. We propose
a neural Modular Task-oriented Dialogue System(MTDS) framework, in which a few
expert bots are combined to generate the response for a given dialogue context.
MTDS consists of a chair bot and several expert bots. Each expert bot is
specialized for a particular situation, e.g., one domain, one type of action of
a system, etc. The chair bot coordinates multiple expert bots and adaptively
selects an expert bot to generate the appropriate response. We further propose
a Token-level Mixture-of-Expert (TokenMoE) model to implement MTDS, where the
expert bots predict multiple tokens at each timestamp and the chair bot
determines the final generated token by fully taking into consideration the
outputs of all expert bots. Both the chair bot and the expert bots are jointly
trained in an end-to-end fashion. To verify the effectiveness of TokenMoE, we
carry out extensive experiments on a benchmark dataset. Compared with the
baseline using a single-module model, our TokenMoE improves the performance by
8.1% of inform rate and 0.8% of success rate.Comment: Proceedings of the 2019 SIGIR Workshop WCIS: Workshop on
Conversational Interaction System
Learning Channel Importance for High Content Imaging with Interpretable Deep Input Channel Mixing
Uncovering novel drug candidates for treating complex diseases remain one of
the most challenging tasks in early discovery research. To tackle this
challenge, biopharma research established a standardized high content imaging
protocol that tags different cellular compartments per image channel. In order
to judge the experimental outcome, the scientist requires knowledge about the
channel importance with respect to a certain phenotype for decoding the
underlying biology. In contrast to traditional image analysis approaches, such
experiments are nowadays preferably analyzed by deep learning based approaches
which, however, lack crucial information about the channel importance. To
overcome this limitation, we present a novel approach which utilizes
multi-spectral information of high content images to interpret a certain aspect
of cellular biology. To this end, we base our method on image blending concepts
with alpha compositing for an arbitrary number of channels. More specifically,
we introduce DCMIX, a lightweight, scaleable and end-to-end trainable mixing
layer which enables interpretable predictions in high content imaging while
retaining the benefits of deep learning based methods. We employ an extensive
set of experiments on both MNIST and RXRX1 datasets, demonstrating that DCMIX
learns the biologically relevant channel importance without scarifying
prediction performance.Comment: Accepted @ DAGM German Conference on Pattern Recognition (GCPR) 202
PhoneMD: Learning to Diagnose Parkinson's Disease from Smartphone Data
Parkinson's disease is a neurodegenerative disease that can affect a person's
movement, speech, dexterity, and cognition. Clinicians primarily diagnose
Parkinson's disease by performing a clinical assessment of symptoms. However,
misdiagnoses are common. One factor that contributes to misdiagnoses is that
the symptoms of Parkinson's disease may not be prominent at the time the
clinical assessment is performed. Here, we present a machine-learning approach
towards distinguishing between people with and without Parkinson's disease
using long-term data from smartphone-based walking, voice, tapping and memory
tests. We demonstrate that our attentive deep-learning models achieve
significant improvements in predictive performance over strong baselines (area
under the receiver operating characteristic curve = 0.85) in data from a cohort
of 1853 participants. We also show that our models identify meaningful features
in the input data. Our results confirm that smartphone data collected over
extended periods of time could in the future potentially be used as a digital
biomarker for the diagnosis of Parkinson's disease.Comment: AAAI Conference on Artificial Intelligence 201
Variable-lag Granger Causality for Time Series Analysis
Granger causality is a fundamental technique for causal inference in time
series data, commonly used in the social and biological sciences. Typical
operationalizations of Granger causality make a strong assumption that every
time point of the effect time series is influenced by a combination of other
time series with a fixed time delay. However, the assumption of the fixed time
delay does not hold in many applications, such as collective behavior,
financial markets, and many natural phenomena. To address this issue, we
develop variable-lag Granger causality, a generalization of Granger causality
that relaxes the assumption of the fixed time delay and allows causes to
influence effects with arbitrary time delays. In addition, we propose a method
for inferring variable-lag Granger causality relations. We demonstrate our
approach on an application for studying coordinated collective behavior and
show that it performs better than several existing methods in both simulated
and real-world datasets. Our approach can be applied in any domain of time
series analysis.Comment: This paper will be appeared in the proceeding of 2019 IEEE
International Conference on Data Science and Advanced Analytics (DSAA). The R
package is available at https://github.com/DarkEyes/VLTimeSeriesCausalit
- ā¦