593,792 research outputs found

    Measuring reproducibility of high-throughput experiments

    Full text link
    Reproducibility is essential to reliable scientific discovery in high-throughput experiments. In this work we propose a unified approach to measure the reproducibility of findings identified from replicate experiments and identify putative discoveries using reproducibility. Unlike the usual scalar measures of reproducibility, our approach creates a curve, which quantitatively assesses when the findings are no longer consistent across replicates. Our curve is fitted by a copula mixture model, from which we derive a quantitative reproducibility score, which we call the "irreproducible discovery rate" (IDR) analogous to the FDR. This score can be computed at each set of paired replicate ranks and permits the principled setting of thresholds both for assessing reproducibility and combining replicates. Since our approach permits an arbitrary scale for each replicate, it provides useful descriptive measures in a wide variety of situations to be explored. We study the performance of the algorithm using simulations and give a heuristic analysis of its theoretical properties. We demonstrate the effectiveness of our method in a ChIP-seq experiment.Comment: Published in at http://dx.doi.org/10.1214/11-AOAS466 the Annals of Applied Statistics (http://www.imstat.org/aoas/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Reproducible Econometric Simulations

    Get PDF
    Reproducibility of economic research has attracted considerable attention in recent years. So far, the discussion has focused on reproducibility of empirical analyses. This paper addresses a further aspect of reproducibility, the reproducibility of computational experiments. We examine the current situation in econometrics and derive a set of guidelines from our findings. To illustrate how computational experiments could be conducted and reported we present an example from time series econometrics that explores the finite-sample power of certain structural change tests.computational experiment, reproducibility, simulation, software.

    Reproducibility of a noisy limit-cycle oscillator induced by a fluctuating input

    Get PDF
    Reproducibility of a noisy limit-cycle oscillator driven by a random piecewise constant signal is analyzed. By reducing the model to random phase maps, it is shown that the reproducibility of the limit cycle generally improves when the phase maps are monotonically increasing.Comment: 4 pages, 3 figures, Prog. Theoret. Phys. Suppl. 200

    Reproducibility of deep inspiration breath hold for prone left-sided whole breast irradiation

    Get PDF
    Background: Investigating reproducibility and instability of deep inspiration breath hold (DIBH) in the prone position to reduce heart dose for left-sided whole breast irradiation. Methods: Thirty patients were included and underwent 2 prone DIBH CT-scans during simulation. Overlap indices were calculated for the ipsilateral breast, heart and lungs to evaluate the anatomical reproducibility of the DIBH maneuver. The breathing motion of 21 patients treated with prone DIBH were registered using magnetic probes. These breathing curves were investigated to gain data on intra-fraction reproducibility and instability of the different DIBH cycles during treatment. Results: Overlap index was 0.98 for the ipsilateral breast and 0.96 for heart and both lungs between the 2 prone DIBH-scans. The magnetic sensors reported population amplitudes of 2.8 +/- 1.3 mm for shallow breathing and 11.7 +/- 4.7 mm for DIBH, an intra-fraction standard deviation of 1.0 +/- 0.4 mm for DIBH, an intra-breath hold instability of 1.0 +/- 0.6 mm and a treatment time of 300 +/- 69 s. Conclusion: Prone DIBH can be accurately clinically implemented with acceptable reproducibility and instability

    Sharing and Preserving Computational Analyses for Posterity with encapsulator

    Get PDF
    Open data and open-source software may be part of the solution to science's "reproducibility crisis", but they are insufficient to guarantee reproducibility. Requiring minimal end-user expertise, encapsulator creates a "time capsule" with reproducible code in a self-contained computational environment. encapsulator provides end-users with a fully-featured desktop environment for reproducible research.Comment: 11 pages, 6 figure

    Predicting computational reproducibility of data analysis pipelines in large population studies using collaborative filtering

    Full text link
    Evaluating the computational reproducibility of data analysis pipelines has become a critical issue. It is, however, a cumbersome process for analyses that involve data from large populations of subjects, due to their computational and storage requirements. We present a method to predict the computational reproducibility of data analysis pipelines in large population studies. We formulate the problem as a collaborative filtering process, with constraints on the construction of the training set. We propose 6 different strategies to build the training set, which we evaluate on 2 datasets, a synthetic one modeling a population with a growing number of subject types, and a real one obtained with neuroinformatics pipelines. Results show that one sampling method, "Random File Numbers (Uniform)" is able to predict computational reproducibility with a good accuracy. We also analyze the relevance of including file and subject biases in the collaborative filtering model. We conclude that the proposed method is able to speedup reproducibility evaluations substantially, with a reduced accuracy loss

    Replicability is not Reproducibility:\ud Nor is it Good Science

    Get PDF
    At various machine learning conferences, at various times, there have been discussions arising from the inability to replicate the experimental results published in a paper. There seems to be a wide spread view that we need to do something to address this problem, as it is essential to the advancement of our field. The most compelling argument would seem to be that reproducibility of experimental results is the hallmark of science. Therefore, given that most of us regard machine learning as a scientific discipline, being able to replicate experiments is paramount. I want to challenge this view by separating the notion of reproducibility, a generally desirable property, from replicability, its poor cousin. I claim there are important differences between the two. Reproducibility requires changes; replicability avoids them. Although reproducibility is desirable, I contend that the impoverished version, replicability, is one not worth having

    Reproducibility of lymphovascular space invasion (LVSI) assessment in endometrial cancer

    Get PDF
    Aims Lymphovascular space invasion (LVSI) in endometrial cancer (EC) is an important prognostic variable impacting on a patient's individual recurrence risk and adjuvant treatment recommendations. Recent work has shown that grading the extent of LVSI further improves its prognostic strength in patients with stage I endometrioid EC. Despite this, there is little information on the reproducibility of LVSI assessment in EC. Therefore, we designed a study to evaluate interobserver agreement in discriminating true LVSI from LVSI mimics (Phase I) and reproducibility of grading extent of LVSI (Phase II). Methods and results Scanned haematoxylin and eosin (H&E) slides of endometrioid EC (EEC) with a predefined possible LVSI focus were hosted on a website and assessed by a panel of six European gynaecological pathologists. In Phase I, 48 H&E slides were included for LVSI assessment and in Phase II, 42 H&E slides for LVSI grading. Each observer was instructed to apply the criteria for LVSI used in daily practice. The degree of agreement was measured using the two-way absolute agreement average-measures intraclass correlation coefficient (ICC). Reproducibility of LVSI assessment (ICC = 0.64, P < 0.001) and LVSI grading (ICC = 0.62, P < 0.001) in EEC was substantial among the observers. Conclusions Given the good reproducibility of LVSI, this study further supports the important role of LVSI in decision algorithms for adjuvant treatment

    Reproducibility Performance Test of Multi Metal Oxide Catalyst in Selective Oxidation of Propane using Combinatorial Technology

    Get PDF
    The concept of rapid catalyst screening using combinatorial technology was applied in the development of selective oxidation catalyst. In this paper, the design, Design of Experiment (DOE) and catalytic results are discussed to demonstrate the importance and versatility of such technology. The instrument is implaced in COMBICAT (Universiti Malaya) to rapidly support parallel testing of catalytic material using continuous fixed bed reactor technology programme. It is used for the automated parallel testing of selective oxidation of propane to acrylic acid over some types of multi metal oxide catalysts. The configuration of the ‘nanoflow’ is shown to be suitable to screen catalytic performance, and its operating conditions were mimicked closely to conventional laboratory as well as to industrial conditions. The results obtained gave very good reproducibility
    corecore