335 research outputs found
Sieve estimation of constant and time-varying coefficients in nonlinear ordinary differential equation models by considering both numerical error and measurement error
This article considers estimation of constant and time-varying coefficients
in nonlinear ordinary differential equation (ODE) models where analytic
closed-form solutions are not available. The numerical solution-based nonlinear
least squares (NLS) estimator is investigated in this study. A numerical
algorithm such as the Runge--Kutta method is used to approximate the ODE
solution. The asymptotic properties are established for the proposed estimators
considering both numerical error and measurement error. The B-spline is used to
approximate the time-varying coefficients, and the corresponding asymptotic
theories in this case are investigated under the framework of the sieve
approach. Our results show that if the maximum step size of the -order
numerical algorithm goes to zero at a rate faster than , the
numerical error is negligible compared to the measurement error. This result
provides a theoretical guidance in selection of the step size for numerical
evaluations of ODEs. Moreover, we have shown that the numerical solution-based
NLS estimator and the sieve NLS estimator are strongly consistent. The sieve
estimator of constant parameters is asymptotically normal with the same
asymptotic co-variance as that of the case where the true ODE solution is
exactly known, while the estimator of the time-varying parameter has the
optimal convergence rate under some regularity conditions. The theoretical
results are also developed for the case when the step size of the ODE numerical
solver does not go to zero fast enough or the numerical error is comparable to
the measurement error. We illustrate our approach with both simulation studies
and clinical data on HIV viral dynamics.Comment: Published in at http://dx.doi.org/10.1214/09-AOS784 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Dynamic Modeling of Sparse Longitudinal Data and Functional Snippets With Stochastic Differential Equations
Sparse functional/longitudinal data have attracted widespread interest due to
the prevalence of such data in social and life sciences. A prominent scenario
where such data are routinely encountered are accelerated longitudinal studies,
where subjects are enrolled in the study at a random time and are only tracked
for a short amount of time relative to the domain of interest. The statistical
analysis of such functional snippets is challenging since information for the
far-off-diagonal regions of the covariance structure is missing. Our main
methodological contribution is to address this challenge by bypassing
covariance estimation and instead modeling the underlying process as the
solution of a data-adaptive stochastic differential equation. Taking advantage
of the interface between Gaussian functional data and stochastic differential
equations makes it possible to efficiently reconstruct the target process by
estimating its dynamic distribution. The proposed approach allows one to
consistently recover forward sample paths from functional snippets at the
subject level. We establish the existence and uniqueness of the solution to the
proposed data-driven stochastic differential equation and derive rates of
convergence for the corresponding estimators. The finite-sample performance is
demonstrated with simulation studies and functional snippets arising from a
growth study and spinal bone mineral density data.Comment: 30 pages, 5 figures, 1 tabl
Dynamic predictions using flexible joint models of longitudinal and time-to-event data.
Joint models for longitudinal and time-to-event data are particularly relevant to many clinical studies where longitudinal biomarkers could be highly associated with a time-to-event outcome. A cutting-edge research direction in this area is dynamic predictions of patient prognosis (e.g., survival probabilities) given all available biomarker information, recently boosted by the stratified/personalized medicine initiative. As these dynamic predictions are individualized, flexible models are desirable in order to appropriately characterize each individual longitudinal trajectory. In this paper, we propose a new joint model using individual-level penalized splines (P-splines) to flexibly characterize the coevolution of the longitudinal and time-to-event processes. An important feature of our approach is that dynamic predictions of the survival probabilities are straightforward as the posterior distribution of the random P-spline coefficients given the observed data is a multivariate skew-normal distribution. The proposed methods are illustrated with data from the HIV Epidemiology Research Study. Our simulation results demonstrate that our model has better dynamic prediction performance than other existing approaches.Medical Research Council. Grant Numbers: G0902100, MR/K014811/1
Unit Programme. Grant Numbers: U105261167, U64-CCU10675
US Centers for Disease Control and Preventio
Application of targeted maximum likelihood estimation in public health and epidemiological studies: a systematic review
BACKGROUND: The Targeted Maximum Likelihood Estimation (TMLE) statistical data analysis framework integrates machine learning, statistical theory, and statistical inference to provide a least biased, efficient and robust strategy for estimation and inference of a variety of statistical and causal parameters. We describe and evaluate the epidemiological applications that have benefited from recent methodological developments. METHODS: We conducted a systematic literature review in PubMed for articles that applied any form of TMLE in observational studies. We summarised the epidemiological discipline, geographical location, expertise of the authors, and TMLE methods over time. We used the Roadmap of Targeted Learning and Causal Inference to extract key methodological aspects of the publications. We showcase the contributions to the literature of these TMLE results. RESULTS: Of the 89 publications included, 33% originated from the University of California at Berkeley, where the framework was first developed by Professor Mark van der Laan. By 2022, 59% of the publications originated from outside the United States and explored up to 7 different epidemiological disciplines in 2021-22. Double-robustness, bias reduction and model misspecification were the main motivations that drew researchers towards the TMLE framework. Through time, a wide variety of methodological, tutorial and software-specific articles were cited, owing to the constant growth of methodological developments around TMLE. CONCLUSIONS: There is a clear dissemination trend of the TMLE framework to various epidemiological disciplines and to increasing numbers of geographical areas. The availability of R packages, publication of tutorial papers, and involvement of methodological experts in applied publications have contributed to an exponential increase in the number of studies that understood the benefits, and adoption, of TMLE
Optimal treatment allocations in space and time for on-line control of an emerging infectious disease
A key component in controlling the spread of an epidemic is deciding where, whenand to whom to apply an intervention.We develop a framework for using data to informthese decisionsin realtime.We formalize a treatment allocation strategy as a sequence of functions, oneper treatment period, that map up-to-date information on the spread of an infectious diseaseto a subset of locations where treatment should be allocated. An optimal allocation strategyoptimizes some cumulative outcome, e.g. the number of uninfected locations, the geographicfootprint of the disease or the cost of the epidemic. Estimation of an optimal allocation strategyfor an emerging infectious disease is challenging because spatial proximity induces interferencebetween locations, the number of possible allocations is exponential in the number oflocations, and because disease dynamics and intervention effectiveness are unknown at outbreak.We derive a Bayesian on-line estimator of the optimal allocation strategy that combinessimulationâoptimization with Thompson sampling.The estimator proposed performs favourablyin simulation experiments. This work is motivated by and illustrated using data on the spread ofwhite nose syndrome, which is a highly fatal infectious disease devastating bat populations inNorth America
Frontiers in Nonparametric Statistics
The goal of this workshop was to discuss recent developments of nonparametric statistical inference. A particular focus was on high dimensional statistics, semiparametrics, adaptation, nonparametric bayesian statistics, shape constraint estimation and statistical inverse problems. The close interaction of these issues with optimization, machine learning and inverse problems has been addressed as well
- âŠ