984 research outputs found
Calibrated Prediction with Covariate Shift via Unsupervised Domain Adaptation
Reliable uncertainty estimates are an important tool for helping autonomous agents or human decision makers understand and leverage predictive models. However, existing approaches to estimating uncertainty largely ignore the possibility of covariate shift—i.e., where the real-world data distribution may differ from the training distribution. As a consequence, existing algorithms can overestimate certainty, possibly yielding a false sense of confidence in the predictive model. We propose an algorithm for calibrating predictions that accounts for the possibility of covariate shift, given labeled examples from the training distribution and unlabeled examples from the real-world distribution. Our algorithm uses importance weighting to correct for the shift from the training to the real-world distribution. However, importance weighting relies on the training and real-world distributions to be sufficiently close. Building on ideas from domain adaptation, we additionally learn a feature map that tries to equalize these two distributions. In an empirical evaluation, we show that our proposed approach outperforms existing approaches to calibrated prediction when there is covariate shift
Graceful Degradation and Related Fields
When machine learning models encounter data which is out of the distribution
on which they were trained they have a tendency to behave poorly, most
prominently over-confidence in erroneous predictions. Such behaviours will have
disastrous effects on real-world machine learning systems. In this field
graceful degradation refers to the optimisation of model performance as it
encounters this out-of-distribution data. This work presents a definition and
discussion of graceful degradation and where it can be applied in deployed
visual systems. Following this a survey of relevant areas is undertaken,
novelly splitting the graceful degradation problem into active and passive
approaches. In passive approaches, graceful degradation is handled and achieved
by the model in a self-contained manner, in active approaches the model is
updated upon encountering epistemic uncertainties. This work communicates the
importance of the problem and aims to prompt the development of machine
learning strategies that are aware of graceful degradation
Uncertainty-Induced Transferability Representation for Source-Free Unsupervised Domain Adaptation
Source-free unsupervised domain adaptation (SFUDA) aims to learn a target
domain model using unlabeled target data and the knowledge of a well-trained
source domain model. Most previous SFUDA works focus on inferring semantics of
target data based on the source knowledge. Without measuring the
transferability of the source knowledge, these methods insufficiently exploit
the source knowledge, and fail to identify the reliability of the inferred
target semantics. However, existing transferability measurements require either
source data or target labels, which are infeasible in SFUDA. To this end,
firstly, we propose a novel Uncertainty-induced Transferability Representation
(UTR), which leverages uncertainty as the tool to analyse the channel-wise
transferability of the source encoder in the absence of the source data and
target labels. The domain-level UTR unravels how transferable the encoder
channels are to the target domain and the instance-level UTR characterizes the
reliability of the inferred target semantics. Secondly, based on the UTR, we
propose a novel Calibrated Adaption Framework (CAF) for SFUDA, including i)the
source knowledge calibration module that guides the target model to learn the
transferable source knowledge and discard the non-transferable one, and ii)the
target semantics calibration module that calibrates the unreliable semantics.
With the help of the calibrated source knowledge and the target semantics, the
model adapts to the target domain safely and ultimately better. We verified the
effectiveness of our method using experimental results and demonstrated that
the proposed method achieves state-of-the-art performances on the three SFUDA
benchmarks. Code is available at https://github.com/SPIresearch/UTR
Confidence Calibration for Systems with Cascaded Predictive Modules
Existing conformal prediction algorithms estimate prediction intervals at
target confidence levels to characterize the performance of a regression model
on new test samples. However, considering an autonomous system consisting of
multiple modules, prediction intervals constructed for individual modules fall
short of accommodating uncertainty propagation over different modules and thus
cannot provide reliable predictions on system behavior. We address this
limitation and present novel solutions based on conformal prediction to provide
prediction intervals calibrated for a predictive system consisting of cascaded
modules (e.g., an upstream feature extraction module and a downstream
regression module). Our key idea is to leverage module-level validation data to
characterize the system-level error distribution without direct access to
end-to-end validation data. We provide theoretical justification and empirical
experimental results to demonstrate the effectiveness of proposed solutions. In
comparison to prediction intervals calibrated for individual modules, our
solutions generate improved intervals with more accurate performance guarantees
for system predictions, which are demonstrated on both synthetic systems and
real-world systems performing overlap prediction for indoor navigation using
the Matterport3D dataset
Uncertainty-based Rejection Wrappers for Black-box Classifiers
Machine Learning as a Service platform is a very sensible choice for practitioners that wantto incorporate machine learning to their products while reducing times and costs. However, to benefit theiradvantages, a method for assessing their performance when applied to a target application is needed. In thiswork, we present a robust uncertainty-based method for evaluating the performance of both probabilistic andcategorical classification black-box models, in particular APIs, that enriches the predictions obtained withan uncertainty score. This uncertainty score enables the detection of inputs with very confident but erroneouspredictions while protecting against out of distribution data points when deploying the model in a productivesetting. We validate the proposal in different natural language processing and computer vision scenarios.Moreover, taking advantage of the computed uncertainty score, we show that one can significantly increasethe robustness and performance of the resulting classification system by rejecting uncertain prediction
Bayesian Nested Latent Class Models for Cause-of-Death Assignment using Verbal Autopsies Across Multiple Domains
Understanding cause-specific mortality rates is crucial for monitoring
population health and designing public health interventions. Worldwide,
two-thirds of deaths do not have a cause assigned. Verbal autopsy (VA) is a
well-established tool to collect information describing deaths outside of
hospitals by conducting surveys to caregivers of a deceased person. It is
routinely implemented in many low- and middle-income countries. Statistical
algorithms to assign cause of death using VAs are typically vulnerable to the
distribution shift between the data used to train the model and the target
population. This presents a major challenge for analyzing VAs as labeled data
are usually unavailable in the target population. This article proposes a
Latent Class model framework for VA data (LCVA) that jointly models VAs
collected over multiple heterogeneous domains, assign cause of death for
out-of-domain observations, and estimate cause-specific mortality fractions for
a new domain. We introduce a parsimonious representation of the joint
distribution of the collected symptoms using nested latent class models and
develop an efficient algorithm for posterior inference. We demonstrate that
LCVA outperforms existing methods in predictive performance and scalability.
Supplementary materials for this article and the R package to implement the
model are available online.Comment: Main paper: 45 pages, 9 figures. Supplement: 20 pages, 16 figures, 2
table
A Saliency-based Clustering Framework for Identifying Aberrant Predictions
In machine learning, classification tasks serve as the cornerstone of a wide
range of real-world applications. Reliable, trustworthy classification is
particularly intricate in biomedical settings, where the ground truth is often
inherently uncertain and relies on high degrees of human expertise for
labeling. Traditional metrics such as precision and recall, while valuable, are
insufficient for capturing the nuances of these ambiguous scenarios. Here we
introduce the concept of aberrant predictions, emphasizing that the nature of
classification errors is as critical as their frequency. We propose a novel,
efficient training methodology aimed at both reducing the misclassification
rate and discerning aberrant predictions. Our framework demonstrates a
substantial improvement in model performance, achieving a 20\% increase in
precision. We apply this methodology to the less-explored domain of veterinary
radiology, where the stakes are high but have not been as extensively studied
compared to human medicine. By focusing on the identification and mitigation of
aberrant predictions, we enhance the utility and trustworthiness of machine
learning classifiers in high-stakes, real-world scenarios, including new
applications in the veterinary world
- …