449 research outputs found
Sufficient Covariate, Propensity Variable and Doubly Robust Estimation
Statistical causal inference from observational studies often requires
adjustment for a possibly multi-dimensional variable, where dimension reduction
is crucial. The propensity score, first introduced by Rosenbaum and Rubin, is a
popular approach to such reduction. We address causal inference within Dawid's
decision-theoretic framework, where it is essential to pay attention to
sufficient covariates and their properties. We examine the role of a propensity
variable in a normal linear model. We investigate both population-based and
sample-based linear regressions, with adjustments for a multivariate covariate
and for a propensity variable. In addition, we study the augmented inverse
probability weighted estimator, involving a combination of a response model and
a propensity model. In a linear regression with homoscedasticity, a propensity
variable is proved to provide the same estimated causal effect as multivariate
adjustment. An estimated propensity variable may, but need not, yield better
precision than the true propensity variable. The augmented inverse probability
weighted estimator is doubly robust and can improve precision if the propensity
model is correctly specified
Confidence Intervals for the Area Under the Receiver Operating Characteristic Curve in the Presence of Ignorable Missing Data
Receiver operating characteristic curves are widely used as a measure of accuracy of diagnostic tests and can be summarised using the area under the receiver operating characteristic curve (AUC). Often, it is useful to construct a confidence interval for the AUC; however, because there are a number of different proposed methods to measure variance of the AUC, there are thus many different resulting methods for constructing these intervals. In this article, we compare different methods of constructing Waldâtype confidence interval in the presence of missing data where the missingness mechanism is ignorable. We find that constructing confidence intervals using multiple imputation based on logistic regression gives the most robust coverage probability and the choice of confidence interval method is less important. However, when missingness rate is less severe (e.g. less than 70%), we recommend using Newcombe\u27s Wald method for constructing confidence intervals along with multiple imputation using predictive mean matching
Within-Household Selection Methods: A Critical Review and Experimental Examination
Probability samples are necessary for making statistical inferences to the general population (Baker et al. 2013). Some countries (e.g. Sweden) have population registers from which to randomly select samples of adults. The U.S. and many other countries, however, do not have population registers. Instead, researchers (i) select a probability sample of households from lists of areas, addresses, or telephone numbers and (ii) select an adult within these sampled households. The process by which individuals are selected from sampled households to obtain a probability-based sample of individuals is called within-household (or within-unit) selection (Gaziano 2005).Within-household selection aims to provide each member of a sampled household with a known, nonzero chance of being selected for the survey (Gaziano 2005; Lavrakas 2008). Thus, it helps to ensure that the sample represents the target population rather than only those most willing and available to participate and, as such, reduces total survey error (TSE).
In interviewer-administered surveys, trained interviewers can implement a prespecified within-household selection procedure, making the selection process relatively straightforward. In self-administered surveys, within-household selection is more challenging because households must carry out the selection task themselves. This can lead to errors in the selection process or nonresponse, resulting in too many or too few of certain types of people in the data (e.g. typically too many female, highly educated, older, and white respondents), and may also lead to biased estimates for other items. We expect the smallest biases in estimates for items that do not differ across household members (e.g. political views, household income) and the largest biases for items that do differ across household members (e.g. household division of labor).
In this chapter, we review recent literature on within-household selection across survey modes, identify the methodological requirements of studying within-household selection methods experimentally, provide an example of an experiment designed to improve the quality of selecting an adult within a household in mail surveys, and summarize current implications for survey practice regarding within-household selection. We focus on selection of one adult out of all possible adults in a household; screening households for members who have particular characteristics has additional complications (e.g. Tourangeau et al. 2012; Brick et al. 2016; Brick et al. 2011), although designing experimental studies for screening follows the same principles
Semiparametric theory and empirical processes in causal inference
In this paper we review important aspects of semiparametric theory and
empirical processes that arise in causal inference problems. We begin with a
brief introduction to the general problem of causal inference, and go on to
discuss estimation and inference for causal effects under semiparametric
models, which allow parts of the data-generating process to be unrestricted if
they are not of particular interest (i.e., nuisance functions). These models
are very useful in causal problems because the outcome process is often complex
and difficult to model, and there may only be information available about the
treatment process (at best). Semiparametric theory gives a framework for
benchmarking efficiency and constructing estimators in such settings. In the
second part of the paper we discuss empirical process theory, which provides
powerful tools for understanding the asymptotic behavior of semiparametric
estimators that depend on flexible nonparametric estimators of nuisance
functions. These tools are crucial for incorporating machine learning and other
modern methods into causal inference analyses. We conclude by examining related
extensions and future directions for work in semiparametric causal inference
Meaningful time for professional growth or a waste of time? A study in five countries on teachersâ experiences within masterâs dissertation/thesis work
The relationship between masterâs thesis work and teachersâ professional development has rarely been explored empirically, yet. Drawing upon a larger study, this paper investigates how teachers who were studying for or who have recently graduated from Master of Education programmes offered in five countries â Poland, Portugal, England, Latvia, Romania â perceive the usefulness of dissertation/thesis work for their professional development and how they attempt to use their MA research results in their (future) teaching practice. Results suggest that although most respondents recognized their MA dissertation/thesis work as having a positive impact on their professional development by enhancing their professionalism, personal development and growth, and understanding the relationship between research and practice, they were less confident about the use of MA research findings in their (future) workplaces. These results are discussed in the context of current challenges regarding masterâs level education for teachers, national governmentsâ educational policies, and the relationship between research, teachersâ practices and professional development
Time-integrated luminosity recorded by the BABAR detector at the PEP-II e+e- collider
This article is the Preprint version of the final published artcile which can be accessed at the link below.We describe a measurement of the time-integrated luminosity of the data collected by the BABAR experiment at the PEP-II asymmetric-energy e+e- collider at the Ï(4S), Ï(3S), and Ï(2S) resonances and in a continuum region below each resonance. We measure the time-integrated luminosity by counting e+e-âe+e- and (for the Ï(4S) only) e+e-âÎŒ+ÎŒ- candidate events, allowing additional photons in the final state. We use data-corrected simulation to determine the cross-sections and reconstruction efficiencies for these processes, as well as the major backgrounds. Due to the large cross-sections of e+e-âe+e- and e+e-âÎŒ+ÎŒ-, the statistical uncertainties of the measurement are substantially smaller than the systematic uncertainties. The dominant systematic uncertainties are due to observed differences between data and simulation, as well as uncertainties on the cross-sections. For data collected on the Ï(3S) and Ï(2S) resonances, an additional uncertainty arises due to Ïâe+e-X background. For data collected off the Ï resonances, we estimate an additional uncertainty due to time dependent efficiency variations, which can affect the short off-resonance runs. The relative uncertainties on the luminosities of the on-resonance (off-resonance) samples are 0.43% (0.43%) for the Ï(4S), 0.58% (0.72%) for the Ï(3S), and 0.68% (0.88%) for the Ï(2S).This work is supported by the US Department of Energy and National Science Foundation, the Natural Sciences and Engineering Research Council (Canada), the Commissariat Ă lâEnergie Atomique and Institut National de Physique NuclĂ©aire et de Physiquedes Particules (France), the Bundesministerium fĂŒr Bildung und Forschung and Deutsche Forschungsgemeinschaft (Germany), the Istituto Nazionale di Fisica Nucleare (Italy), the Foundation for Fundamental Research on Matter (The Netherlands), the Research Council of Norway, the Ministry of Education and Science of the Russian Federation, Ministerio de Ciencia e InnovaciĂłn (Spain), and the Science and Technology Facilities Council (United Kingdom). Individuals have received support from the Marie-Curie IEF program (European Union) and the A.P. Sloan Foundation (USA)
- âŠ