855 research outputs found

    A simulation model for diarrhoea and other common recurrent infections: a tool for exploring epidemiological methods

    Get PDF
    The measurement and analysis of common recurrent conditions such as diarrhoea, respiratory infections or fever pose methodological challenges with regard to case definition, disease surveillance and statistical analysis. In this paper we describe a flexible and robust model that can generate simulated longitudinal datasets for a range of recurrent infections, reflecting the stochastic processes that underpin the data collected in the field. It can be used to evaluate and compare alternative disease definitions, surveillance strategies and statistical methods under ‘controlled conditions’. Parameters in the model include: characterizing the distributions of the individual disease incidence and the duration of disease episodes; allowing the average disease duration to depend on an individual's number of episodes (simulating a correlation between incidence and duration); making the individual risk of disease depend on the occurrence of previous episodes (simulating autocorrelation of successive episodes); finally, incorporating seasonal variation of disease

    Data production models for the CDF experiment

    Get PDF
    The data production for the CDF experiment is conducted on a large Linux PC farm designed to meet the needs of data collection at a maximum rate of 40 MByte/sec. We present two data production models that exploits advances in computing and communication technology. The first production farm is a centralized system that has achieved a stable data processing rate of approximately 2 TByte per day. The recently upgraded farm is migrated to the SAM (Sequential Access to data via Metadata) data handling system. The software and hardware of the CDF production farms has been successful in providing large computing and data throughput capacity to the experiment.Comment: 8 pages, 9 figures; presented at HPC Asia2005, Beijing, China, Nov 30 - Dec 3, 200

    Novel statistical approaches for non-normal censored immunological data: analysis of cytokine and gene expression data

    Get PDF
    Background: For several immune-mediated diseases, immunological analysis will become more complex in the future with datasets in which cytokine and gene expression data play a major role. These data have certain characteristics that require sophisticated statistical analysis such as strategies for non-normal distribution and censoring. Additionally, complex and multiple immunological relationships need to be adjusted for potential confounding and interaction effects. Objective: We aimed to introduce and apply different methods for statistical analysis of non-normal censored cytokine and gene expression data. Furthermore, we assessed the performance and accuracy of a novel regression approach in order to allow adjusting for covariates and potential confounding. Methods: For non-normally distributed censored data traditional means such as the Kaplan-Meier method or the generalized Wilcoxon test are described. In order to adjust for covariates the novel approach named Tobit regression on ranks was introduced. Its performance and accuracy for analysis of non-normal censored cytokine/gene expression data was evaluated by a simulation study and a statistical experiment applying permutation and bootstrapping. Results: If adjustment for covariates is not necessary traditional statistical methods are adequate for non-normal censored data. Comparable with these and appropriate if additional adjustment is required, Tobit regression on ranks is a valid method. Its power, type-I error rate and accuracy were comparable to the classical Tobit regression. Conclusion: Non-normally distributed censored immunological data require appropriate statistical methods. Tobit regression on ranks meets these requirements and can be used for adjustment for covariates and potential confounding in large and complex immunological datasets

    Data processing model for the CDF experiment

    Get PDF
    The data processing model for the CDF experiment is described. Data processing reconstructs events from parallel data streams taken with different combinations of physics event triggers and further splits the events into datasets of specialized physics datasets. The design of the processing control system faces strict requirements on bookkeeping records, which trace the status of data files and event contents during processing and storage. The computing architecture was updated to meet the mass data flow of the Run II data collection, recently upgraded to a maximum rate of 40 MByte/sec. The data processing facility consists of a large cluster of Linux computers with data movement managed by the CDF data handling system to a multi-petaByte Enstore tape library. The latest processing cycle has achieved a stable speed of 35 MByte/sec (3 TByte/day). It can be readily scaled by increasing CPU and data-handling capacity as required.Comment: 12 pages, 10 figures, submitted to IEEE-TN

    Applied immuno-epidemiological research: an approach for integrating existing knowledge into the statistical analysis of multiple immune markers.

    Get PDF
    BACKGROUND: Immunologists often measure several correlated immunological markers, such as concentrations of different cytokines produced by different immune cells and/or measured under different conditions, to draw insights from complex immunological mechanisms. Although there have been recent methodological efforts to improve the statistical analysis of immunological data, a framework is still needed for the simultaneous analysis of multiple, often correlated, immune markers. This framework would allow the immunologists' hypotheses about the underlying biological mechanisms to be integrated. RESULTS: We present an analytical approach for statistical analysis of correlated immune markers, such as those commonly collected in modern immuno-epidemiological studies. We demonstrate i) how to deal with interdependencies among multiple measurements of the same immune marker, ii) how to analyse association patterns among different markers, iii) how to aggregate different measures and/or markers to immunological summary scores, iv) how to model the inter-relationships among these scores, and v) how to use these scores in epidemiological association analyses. We illustrate the application of our approach to multiple cytokine measurements from 818 children enrolled in a large immuno-epidemiological study (SCAALA Salvador), which aimed to quantify the major immunological mechanisms underlying atopic diseases or asthma. We demonstrate how to aggregate systematically the information captured in multiple cytokine measurements to immunological summary scores aimed at reflecting the presumed underlying immunological mechanisms (Th1/Th2 balance and immune regulatory network). We show how these aggregated immune scores can be used as predictors in regression models with outcomes of immunological studies (e.g. specific IgE) and compare the results to those obtained by a traditional multivariate regression approach. CONCLUSION: The proposed analytical approach may be especially useful to quantify complex immune responses in immuno-epidemiological studies, where investigators examine the relationship among epidemiological patterns, immune response, and disease outcomes

    Recent diarrhoeal illness and risk of lower respiratory infections in children under the age of 5 years

    Get PDF
    Background Children in low-income settings suffering from frequent diarrhoea episodes are also at a high risk of acute lower respiratory infections (ALRI). We explored whether this is due to common risk factors for both conditions or whether diarrhoea can increase the risk of ALRI directly

    Sampling strategies to measure the prevalence of common recurrent infections in longitudinal studies

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Measuring recurrent infections such as diarrhoea or respiratory infections in epidemiological studies is a methodological challenge. Problems in measuring the incidence of recurrent infections include the episode definition, recall error, and the logistics of close follow up. Longitudinal prevalence (LP), the proportion-of-time-ill estimated by repeated prevalence measurements, is an alternative measure to incidence of recurrent infections. In contrast to incidence which usually requires continuous sampling, LP can be measured at intervals. This study explored how many more participants are needed for infrequent sampling to achieve the same study power as frequent sampling.</p> <p>Methods</p> <p>We developed a set of four empirical simulation models representing low and high risk settings with short or long episode durations. The model was used to evaluate different sampling strategies with different assumptions on recall period and recall error.</p> <p>Results</p> <p>The model identified three major factors that influence sampling strategies: (1) the clustering of episodes in individuals; (2) the duration of episodes; (3) the positive correlation between an individual's disease incidence and episode duration. Intermittent sampling (e.g. 12 times per year) often requires only a slightly larger sample size compared to continuous sampling, especially in cluster-randomized trials. The collection of period prevalence data can lead to highly biased effect estimates if the exposure variable is associated with episode duration. To maximize study power, recall periods of 3 to 7 days may be preferable over shorter periods, even if this leads to inaccuracy in the prevalence estimates.</p> <p>Conclusion</p> <p>Choosing the optimal approach to measure recurrent infections in epidemiological studies depends on the setting, the study objectives, study design and budget constraints. Sampling at intervals can contribute to making epidemiological studies and trials more efficient, valid and cost-effective.</p

    A Quasi-Model-Independent Search for New Physics at Large Transverse Momentum

    Get PDF
    We apply a quasi-model-independent strategy ("Sleuth") to search for new high p_T physics in approximately 100 pb^-1 of ppbar collisions at sqrt(s) = 1.8 TeV collected by the DZero experiment during 1992-1996 at the Fermilab Tevatron. Over thirty-two e mu X, W+jets-like, Z+jets-like, and 3(lepton/photon)X exclusive final states are systematically analyzed for hints of physics beyond the standard model. Simultaneous sensitivity to a variety of models predicting new phenomena at the electroweak scale is demonstrated by testing the method on a particular signature in each set of final states. No evidence of new high p_T physics is observed in the course of this search, and we find that 89% of an ensemble of hypothetical similar experimental runs would have produced a final state with a candidate signal more interesting than the most interesting observed in these data.Comment: 28 pages, 17 figures. Submitted to Physical Review

    A measurement of the W boson mass using large rapidity electrons

    Get PDF
    We present a measurement of the W boson mass using data collected by the D0 experiment at the Fermilab Tevatron during 1994--1995. We identify W bosons by their decays to e-nu final states where the electron is detected in a forward calorimeter. We extract the W boson mass, Mw, by fitting the transverse mass and transverse electron and neutrino momentum spectra from a sample of 11,089 W -> e nu decay candidates. We use a sample of 1,687 dielectron events, mostly due to Z -> ee decays, to constrain our model of the detector response. Using the forward calorimeter data, we measure Mw = 80.691 +- 0.227 GeV. Combining the forward calorimeter measurements with our previously published central calorimeter results, we obtain Mw = 80.482 +- 0.091 GeV

    Spin Correlation in tt-bar Production from pp-bar Collisions at sqrt(s)=1.8 TeV

    Get PDF
    The D0 collaboration has performed a study of spin correlation in tt-bar production for the process tt-bar to bb-bar W^+W^-, where the W bosons decay to e-nu or mu-nu. A sample of six events was collected during an exposure of the D0 detector to an integrated luminosity of approximately 125 pb^-1 of sqrt{s}=1.8 TeV pp-bar collisions. The standard model (SM) predicts that the short lifetime of the top quark ensures the transmission of any spin information at production to the tt-bar decay products. The degree of spin correlation is characterized by a correlation coefficient k. We find that k>-0.25 at the 68% confidence level, in agreement with the SM prediction of k=0.88.Comment: Submitted to PRL, Added references, minor changes to tex
    corecore