159 research outputs found
Distance-based Analysis of Machine Learning Prediction Reliability for Datasets in Materials Science and Other Fields
Despite successful use in a wide variety of disciplines for data analysis and
prediction, machine learning (ML) methods suffer from a lack of understanding
of the reliability of predictions due to the lack of transparency and black-box
nature of ML models. In materials science and other fields, typical ML model
results include a significant number of low-quality predictions. This problem
is known to be particularly acute for target systems which differ significantly
from the data used for ML model training. However, to date, a general method
for characterization of the difference between the predicted and training
system has not been available. Here, we show that a simple metric based on
Euclidean feature space distance and sampling density allows effective
separation of the accurately predicted data points from data points with poor
prediction accuracy. We show that the metric effectiveness is enhanced by the
decorrelation of the features using Gram-Schmidt orthogonalization. To
demonstrate the generality of the method, we apply it to support vector
regression models for various small data sets in materials science and other
fields. Our method is computationally simple, can be used with any ML learning
method and enables analysis of the sources of the ML prediction errors.
Therefore, it is suitable for use as a standard technique for the estimation of
ML prediction reliability for small data sets and as a tool for data set
design
The audio auditor: user-level membership inference in Internet of Things voice services
With the rapid development of deep learning techniques, the popularity of voice services implemented on various Internet of Things (IoT) devices is ever increasing. In this paper, we examine user-level membership inference in the problem space of voice services, by designing an audio auditor to verify whether a specific user had unwillingly contributed audio used to train an automatic speech recognition (ASR) model under strict black-box access. With user representation of the input audio data and their corresponding translated text, our trained auditor is effective in user-level audit. We also observe that the auditor trained on specific data can be generalized well regardless of the ASR model architecture. We validate the auditor on ASR models trained with LSTM, RNNs, and GRU algorithms on two state-of-the-art pipelines, the hybrid ASR system and the end-to-end ASR system. Finally, we conduct a real-world trial of our auditor on iPhone Siri, achieving an overall accuracy exceeding 80%. We hope the methodology developed in this paper and findings can inform privacy advocates to overhaul IoT privacy
A manifesto on explainability for artificial intelligence in medicine.
The rapid increase of interest in, and use of, artificial intelligence (AI) in computer applications has raised a parallel concern about its ability (or lack thereof) to provide understandable, or explainable, output to users. This concern is especially legitimate in biomedical contexts, where patient safety is of paramount importance. This position paper brings together seven researchers working in the field with different roles and perspectives, to explore in depth the concept of explainable AI, or XAI, offering a functional definition and conceptual framework or model that can be used when considering XAI. This is followed by a series of desiderata for attaining explainability in AI, each of which touches upon a key domain in biomedicine
- …