49 research outputs found
Generalization error for decision problems
In this entry we review the generalization error for classification and
single-stage decision problems. We distinguish three alternative definitions of
the generalization error which have, at times, been conflated in the statistics
literature and show that these definitions need not be equivalent even
asymptotically. Because the generalization error is a non-smooth functional of
the underlying generative model, standard asymptotic approximations, e.g., the
bootstrap or normal approximations, cannot guarantee correct frequentist
operating characteristics without modification. We provide simple data-adaptive
procedures that can be used to construct asymptotically valid confidence sets
for the generalization error. We conclude the entry with a discussion of
extensions and related problems
Small Sample Inference for Generalization Error in Classification Using the CUD Bound
Confidence measures for the generalization error are crucial when small
training samples are used to construct classifiers. A common approach is to
estimate the generalization error by resampling and then assume the resampled
estimator follows a known distribution to form a confidence set [Kohavi 1995,
Martin 1996,Yang 2006]. Alternatively, one might bootstrap the resampled
estimator of the generalization error to form a confidence set. Unfortunately,
these methods do not reliably provide sets of the desired confidence. The poor
performance appears to be due to the lack of smoothness of the generalization
error as a function of the learned classifier. This results in a non-normal
distribution of the estimated generalization error. We construct a confidence
set for the generalization error by use of a smooth upper bound on the
deviation between the resampled estimate and generalization error. The
confidence set is formed by bootstrapping this upper bound. In cases in which
the approximation class for the classifier can be represented as a parametric
additive model, we provide a computationally efficient algorithm. This method
exhibits superior performance across a series of test and simulated data sets.Comment: Appears in Proceedings of the Twenty-Fourth Conference on Uncertainty
in Artificial Intelligence (UAI2008
An imputation method for estimating the learning curve in classification problems
The learning curve expresses the error rate of a predictive modeling
procedure as a function of the sample size of the training dataset. It
typically is a decreasing, convex function with a positive limiting value. An
estimate of the learning curve can be used to assess whether a modeling
procedure should be expected to become substantially more accurate if
additional training data become available. This article proposes a new
procedure for estimating learning curves using imputation. We focus on
classification, although the idea is applicable to other predictive modeling
settings. Simulation studies indicate that the learning curve can be estimated
with useful accuracy for a roughly four-fold increase in the size of the
training set relative to the available data, and that the proposed imputation
approach outperforms an alternative estimation approach based on parameterizing
the learning curve. We illustrate the method with an application that predicts
the risk of disease progression for people with chronic lymphocytic leukemia
Sufficient Markov Decision Processes with Alternating Deep Neural Networks
Advances in mobile computing technologies have made it possible to monitor
and apply data-driven interventions across complex systems in real time. Markov
decision processes (MDPs) are the primary model for sequential decision
problems with a large or indefinite time horizon. Choosing a representation of
the underlying decision process that is both Markov and low-dimensional is
non-trivial. We propose a method for constructing a low-dimensional
representation of the original decision process for which: 1. the MDP model
holds; 2. a decision strategy that maximizes mean utility when applied to the
low-dimensional representation also maximizes mean utility when applied to the
original process. We use a deep neural network to define a class of potential
process representations and estimate the process of lowest dimension within
this class. The method is illustrated using data from a mobile study on heavy
drinking and smoking among college students.Comment: 31 pages, 3 figures, extended abstract in the proceedings of
RLDM2017. (v2 revisions: Fixed a minor bug in the code w.r.t. setting seed,
as a result numbers in the simulation experiments had some slight changes,
but conclusions stayed the same. Corrected typos. Improved notations.
Hierarchical Continuous Time Hidden Markov Model, with Application in Zero-Inflated Accelerometer Data
Wearable devices including accelerometers are increasingly being used to
collect high-frequency human activity data in situ. There is tremendous
potential to use such data to inform medical decision making and public health
policies. However, modeling such data is challenging as they are
high-dimensional, heterogeneous, and subject to informative missingness, e.g.,
zero readings when the device is removed by the participant. We propose a
flexible and extensible continuous-time hidden Markov model to extract
meaningful activity patterns from human accelerometer data. To facilitate
estimation with massive data we derive an efficient learning algorithm that
exploits the hierarchical structure of the parameters indexing the proposed
model. We also propose a bootstrap procedure for interval estimation. The
proposed methods are illustrated using data from the 2003 - 2004 and 2005 -
2006 National Health and Nutrition Examination Survey.Comment: 18 pages, 4 figure
Interpretable Dynamic Treatment Regimes
Precision medicine is currently a topic of great interest in clinical and
intervention science. One way to formalize precision medicine is through a
treatment regime, which is a sequence of decision rules, one per stage of
clinical intervention, that map up-to-date patient information to a recommended
treatment. An optimal treatment regime is defined as maximizing the mean of
some cumulative clinical outcome if applied to a population of interest. It is
well-known that even under simple generative models an optimal treatment regime
can be a highly nonlinear function of patient information. Consequently, a
focal point of recent methodological research has been the development of
flexible models for estimating optimal treatment regimes. However, in many
settings, estimation of an optimal treatment regime is an exploratory analysis
intended to generate new hypotheses for subsequent research and not to directly
dictate treatment to new patients. In such settings, an estimated treatment
regime that is interpretable in a domain context may be of greater value than
an unintelligible treatment regime built using "black-box" estimation methods.
We propose an estimator of an optimal treatment regime composed of a sequence
of decision rules, each expressible as a list of "if-then" statements that can
be presented as either a paragraph or as a simple flowchart that is immediately
interpretable to domain experts. The discreteness of these lists precludes
smooth, i.e., gradient-based, methods of estimation and leads to non-standard
asymptotics. Nevertheless, we provide a computationally efficient estimation
algorithm, prove consistency of the proposed estimator, and derive rates of
convergence. We illustrate the proposed methods using a series of simulation
examples and application to data from a sequential clinical trial on bipolar
disorder
Using Decision Lists to Construct Interpretable and Parsimonious Treatment Regimes
A treatment regime formalizes personalized medicine as a function from
individual patient characteristics to a recommended treatment. A high-quality
treatment regime can improve patient outcomes while reducing cost, resource
consumption, and treatment burden. Thus, there is tremendous interest in
estimating treatment regimes from observational and randomized studies.
However, the development of treatment regimes for application in clinical
practice requires the long-term, joint effort of statisticians and clinical
scientists. In this collaborative process, the statistician must integrate
clinical science into the statistical models underlying a treatment regime and
the clinician must scrutinize the estimated treatment regime for scientific
validity. To facilitate meaningful information exchange, it is important that
estimated treatment regimes be interpretable in a subject-matter context. We
propose a simple, yet flexible class of treatment regimes whose members are
representable as a short list of if-then statements. Regimes in this class are
immediately interpretable and are therefore an appealing choice for broad
application in practice. We derive a robust estimator of the optimal regime
within this class and demonstrate its finite sample performance using
simulation experiments. The proposed method is illustrated with data from two
clinical trials
Estimation and Optimization of Composite Outcomes
There is tremendous interest in precision medicine as a means to improve
patient outcomes by tailoring treatment to individual characteristics. An
individualized treatment rule formalizes precision medicine as a map from
patient information to a recommended treatment. A treatment rule is defined to
be optimal if it maximizes the mean of a scalar outcome in a population of
interest, e.g., symptom reduction. However, clinical and intervention
scientists often must balance multiple and possibly competing outcomes, e.g.,
symptom reduction and the risk of an adverse event. One approach to precision
medicine in this setting is to elicit a composite outcome which balances all
competing outcomes; unfortunately, eliciting a composite outcome directly from
patients is difficult without a high-quality instrument, and an expert-derived
composite outcome may not account for heterogeneity in patient preferences. We
propose a new paradigm for the study of precision medicine using observational
data that relies solely on the assumption that clinicians are approximately
(i.e., imperfectly) making decisions to maximize individual patient utility.
Estimated composite outcomes are subsequently used to construct an estimator of
an individualized treatment rule which maximizes the mean of patient-specific
composite outcomes. The estimated composite outcomes and estimated optimal
individualized treatment rule provide new insights into patient preference
heterogeneity, clinician behavior, and the value of precision medicine in a
given domain. We derive inference procedures for the proposed estimators under
mild conditions and demonstrate their finite sample performance through a suite
of simulation experiments and an illustrative application to data from a study
of bipolar depression
Thompson Sampling for Pursuit-Evasion Problems
Pursuit-evasion is a multi-agent sequential decision problem wherein a group
of agents known as pursuers coordinate their traversal of a spatial domain to
locate an agent trying to evade them. Pursuit evasion problems arise in a
number of import application domains including defense and route planning.
Learning to optimally coordinate pursuer behaviors so as to minimize time to
capture of the evader is challenging because of a large action space and sparse
noisy state information; consequently, previous approaches have relied
primarily on heuristics. We propose a variant of Thompson Sampling for
pursuit-evasion that allows for the application of existing model-based
planning algorithms. This approach is general in that it allows for an
arbitrary number of pursuers, a general spatial domain, and the integration of
auxiliary information provided by informants. In a suite of simulation
experiments, Thompson Sampling for pursuit evasion significantly reduces
time-to-capture relative to competing algorithms
Bayesian Nonparametric Policy Search with Application to Periodontal Recall Intervals
Tooth loss from periodontal disease is a major public health burden in the
United States. Standard clinical practice is to recommend a dental visit every
six months; however, this practice is not evidence-based, and poor dental
outcomes and increasing dental insurance premiums indicate room for
improvement. We consider a tailored approach that recommends recall time based
on patient characteristics and medical history to minimize disease progression
without increasing resource expenditures. We formalize this method as a dynamic
treatment regime which comprises a sequence of decisions, one per stage of
intervention, that follow a decision rule which maps current patient
information to a recommendation for their next visit time. The dynamics of
periodontal health, visit frequency, and patient compliance are complex, yet
the estimated optimal regime must be interpretable to domain experts if it is
to be integrated into clinical practice. We combine non-parametric Bayesian
dynamics modeling with policy-search algorithms to estimate the optimal dynamic
treatment regime within an interpretable class of regimes. Both simulation
experiments and application to a rich database of electronic dental records
from the HealthPartners HMO shows that our proposed method leads to better
dental health without increasing the average recommended recall time relative
to competing methods