17,595 research outputs found
Causal inference for continuous-time processes when covariates are observed only at discrete times
Most of the work on the structural nested model and g-estimation for causal
inference in longitudinal data assumes a discrete-time underlying data
generating process. However, in some observational studies, it is more
reasonable to assume that the data are generated from a continuous-time process
and are only observable at discrete time points. When these circumstances
arise, the sequential randomization assumption in the observed discrete-time
data, which is essential in justifying discrete-time g-estimation, may not be
reasonable. Under a deterministic model, we discuss other useful assumptions
that guarantee the consistency of discrete-time g-estimation. In more general
cases, when those assumptions are violated, we propose a controlling-the-future
method that performs at least as well as g-estimation in most scenarios and
which provides consistent estimation in some cases where g-estimation is
severely inconsistent. We apply the methods discussed in this paper to
simulated data, as well as to a data set collected following a massive flood in
Bangladesh, estimating the effect of diarrhea on children's height. Results
from different methods are compared in both simulation and the real
application.Comment: Published in at http://dx.doi.org/10.1214/10-AOS830 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Defining and Estimating Intervention Effects for Groups that will Develop an Auxiliary Outcome
It has recently become popular to define treatment effects for subsets of the
target population characterized by variables not observable at the time a
treatment decision is made. Characterizing and estimating such treatment
effects is tricky; the most popular but naive approach inappropriately adjusts
for variables affected by treatment and so is biased. We consider several
appropriate ways to formalize the effects: principal stratification,
stratification on a single potential auxiliary variable, stratification on an
observed auxiliary variable and stratification on expected levels of auxiliary
variables. We then outline identifying assumptions for each type of estimand.
We evaluate the utility of these estimands and estimation procedures for
decision making and understanding causal processes, contrasting them with the
concepts of direct and indirect effects. We motivate our development with
examples from nephrology and cancer screening, and use simulated data and real
data on cancer screening to illustrate the estimation methods.Comment: Published at http://dx.doi.org/10.1214/088342306000000655 in the
Statistical Science (http://www.imstat.org/sts/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Estimating the distribution of dynamic invariants: illustrated with an application to human photo-plethysmographic time series
Dynamic invariants are often estimated from experimental time series with the aim of differentiating between different physical states in the underlying system. The most popular schemes for estimating dynamic invariants are capable of estimating confidence intervals, however, such confidence intervals do not reflect variability in the underlying dynamics. We propose a surrogate based method to estimate the expected distribution of values under the null hypothesis that the underlying deterministic dynamics are stationary. We demonstrate the application of this method by considering four recordings of human pulse waveforms in differing physiological states and show that correlation dimension and entropy are insufficient to differentiate between these states. In contrast, algorithmic complexity can clearly differentiate between all four rhythms
Detecting periodicity in experimental data using linear modeling techniques
Fourier spectral estimates and, to a lesser extent, the autocorrelation
function are the primary tools to detect periodicities in experimental data in
the physical and biological sciences. We propose a new method which is more
reliable than traditional techniques, and is able to make clear identification
of periodic behavior when traditional techniques do not. This technique is
based on an information theoretic reduction of linear (autoregressive) models
so that only the essential features of an autoregressive model are retained.
These models we call reduced autoregressive models (RARM). The essential
features of reduced autoregressive models include any periodicity present in
the data. We provide theoretical and numerical evidence from both experimental
and artificial data, to demonstrate that this technique will reliably detect
periodicities if and only if they are present in the data. There are strong
information theoretic arguments to support the statement that RARM detects
periodicities if they are present. Surrogate data techniques are used to ensure
the converse. Furthermore, our calculations demonstrate that RARM is more
robust, more accurate, and more sensitive, than traditional spectral
techniques.Comment: 10 pages (revtex) and 6 figures. To appear in Phys Rev E. Modified
styl
Identifying the starting point of a spreading process in complex networks
When dealing with the dissemination of epidemics, one important question that
can be asked is the location where the contamination began. In this paper, we
analyze three spreading schemes and propose and validate an effective
methodology for the identification of the source nodes. The method is based on
the calculation of the centrality of the nodes on the sampled network,
expressed here by degree, betweenness, closeness and eigenvector centrality. We
show that the source node tends to have the highest measurement values. The
potential of the methodology is illustrated with respect to three theoretical
complex network models as well as a real-world network, the email network of
the University Rovira i Virgili
Dynamical modeling of collective behavior from pigeon flight data: flock cohesion and dispersion
Several models of flocking have been promoted based on simulations with
qualitatively naturalistic behavior. In this paper we provide the first direct
application of computational modeling methods to infer flocking behavior from
experimental field data. We show that this approach is able to infer general
rules for interaction, or lack of interaction, among members of a flock or,
more generally, any community. Using experimental field measurements of homing
pigeons in flight we demonstrate the existence of a basic distance dependent
attraction/repulsion relationship and show that this rule is sufficient to
explain collective behavior observed in nature. Positional data of individuals
over time are used as input data to a computational algorithm capable of
building complex nonlinear functions that can represent the system behavior.
Topological nearest neighbor interactions are considered to characterize the
components within this model. The efficacy of this method is demonstrated with
simulated noisy data generated from the classical (two dimensional) Vicsek
model. When applied to experimental data from homing pigeon flights we show
that the more complex three dimensional models are capable of predicting and
simulating trajectories, as well as exhibiting realistic collective dynamics.
The simulations of the reconstructed models are used to extract properties of
the collective behavior in pigeons, and how it is affected by changing the
initial conditions of the system. Our results demonstrate that this approach
may be applied to construct models capable of simulating trajectories and
collective dynamics using experimental field measurements of herd movement.
From these models, the behavior of the individual agents (animals) may be
inferred
- …