54 research outputs found

    Using simulation to interpret a discrete time survival model in a complex biological system: fertility and lameness in dairy cows

    Get PDF
    The ever-growing volume of data routinely collected and stored in everyday life presents researchers with a number of opportunities to gain insight and make predictions. This study aimed to demonstrate the usefulness in a specific clinical context of a simulation-based technique called probabilistic sensitivity analysis (PSA) in interpreting the results of a discrete time survival model based on a large dataset of routinely collected dairy herd management data. Data from 12,515 dairy cows (from 39 herds) were used to construct a multilevel discrete time survival model in which the outcome was the probability of a cow becoming pregnant during a given two day period of risk, and presence or absence of a recorded lameness event during various time frames relative to the risk period amongst the potential explanatory variables. A separate simulation model was then constructed to evaluate the wider clinical implications of the model results (i.e. the potential for a herd’s incidence rate of lameness to influence its overall reproductive performance) using PSA. Although the discrete time survival analysis revealed some relatively large associations between lameness events and risk of pregnancy (for example, occurrence of a lameness case within 14 days of a risk period was associated with a 25% reduction in the risk of the cow becoming pregnant during that risk period), PSA revealed that, when viewed in the context of a realistic clinical situation, a herd’s lameness incidence rate is highly unlikely to influence its overall reproductive performance to a meaningful extent in the vast majority of situations. Construction of a simulation model within a PSA framework proved to be a very useful additional step to aid contextualisation of the results from a discrete time survival model, especially where the research is designed to guide on-farm management decisions at population (i.e. herd) rather than individual level

    Uniform Approximation Is More Appropriate for Wilcoxon Rank-Sum Test in Gene Set Analysis

    Get PDF
    Gene set analysis is widely used to facilitate biological interpretations in the analyses of differential expression from high throughput profiling data. Wilcoxon Rank-Sum (WRS) test is one of the commonly used methods in gene set enrichment analysis. It compares the ranks of genes in a gene set against those of genes outside the gene set. This method is easy to implement and it eliminates the dichotomization of genes into significant and non-significant in a competitive hypothesis testing. Due to the large number of genes being examined, it is impractical to calculate the exact null distribution for the WRS test. Therefore, the normal distribution is commonly used as an approximation. However, as we demonstrate in this paper, the normal approximation is problematic when a gene set with relative small number of genes is tested against the large number of genes in the complementary set. In this situation, a uniform approximation is substantially more powerful, more accurate, and less intensive in computation. We demonstrate the advantage of the uniform approximations in Gene Ontology (GO) term analysis using simulations and real data sets

    Enhanced secondary analysis of survival data: reconstructing the data from published Kaplan-Meier survival curves

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>The results of Randomized Controlled Trials (RCTs) on time-to-event outcomes that are usually reported are median time to events and Cox Hazard Ratio. These do not constitute the sufficient statistics required for meta-analysis or cost-effectiveness analysis, and their use in secondary analyses requires strong assumptions that may not have been adequately tested. In order to enhance the quality of secondary data analyses, we propose a method which derives from the published Kaplan Meier survival curves a close approximation to the original individual patient time-to-event data from which they were generated.</p> <p>Methods</p> <p>We develop an algorithm that maps from digitised curves back to KM data by finding numerical solutions to the inverted KM equations, using where available information on number of events and numbers at risk. The reproducibility and accuracy of survival probabilities, median survival times and hazard ratios based on reconstructed KM data was assessed by comparing published statistics (survival probabilities, medians and hazard ratios) with statistics based on repeated reconstructions by multiple observers.</p> <p>Results</p> <p>The validation exercise established there was no material systematic error and that there was a high degree of reproducibility for all statistics. Accuracy was excellent for survival probabilities and medians, for hazard ratios reasonable accuracy can only be obtained if at least numbers at risk or total number of events are reported.</p> <p>Conclusion</p> <p>The algorithm is a reliable tool for meta-analysis and cost-effectiveness analyses of RCTs reporting time-to-event data. It is recommended that all RCTs should report information on numbers at risk and total number of events alongside KM curves.</p

    Effect of COMBinAtion therapy with remote ischemic conditioning and exenatide on the Myocardial Infarct size: a two-by-two factorial randomized trial (COMBAT-MI)

    Get PDF
    Remote ischemic conditioning (RIC) and the GLP-1 analog exenatide activate different cardioprotective pathways and may have additive effects on infarct size (IS). Here, we aimed to assess the efficacy of RIC as compared with sham procedure, and of exenatide, as compared with placebo, and the interaction between both, to reduce IS in humans. We designed a two-by-two factorial, randomized controlled, blinded, multicenter, clinical trial. Patients with ST-segment elevation myocardial infarction receiving primary percutaneous coronary intervention (PPCI) within 6 h of symptoms were randomized to RIC or sham procedure and exenatide or matching placebo. The primary outcome was IS measured by late gadolinium enhancement in cardiac magnetic resonance performed 3–7 days after PPCI. The secondary outcomes were myocardial salvage index, transmurality index, left ventricular ejection fraction and relative microvascular obstruction volume. A total of 378 patients were randomly allocated, and after applying exclusion criteria, 222 patients were available for analysis. There were no significant interactions between the two randomization factors on the primary or secondary outcomes. IS was similar between groups for the RIC (24 ± 11.8% in the RIC group vs 23.7 ± 10.9% in the sham group, P = 0.827) and the exenatide hypotheses (25.1 ± 11.5% in the exenatide group vs 22.5 ± 10.9% in the placebo group, P = 0.092). There were no effects with either RIC or exenatide on the secondary outcomes. Unexpected adverse events or side effects of RIC and exenatide were not observed. In conclusion, neither RIC nor exenatide, or its combination, were able to reduce IS in STEMI patients when administered as an adjunct to PPCI

    Variables with time-varying effects and the Cox model: Some statistical concepts illustrated with a prognostic factor study in breast cancer

    Get PDF
    International audienceBACKGROUND: The Cox model relies on the proportional hazards (PH) assumption, implying that the factors investigated have a constant impact on the hazard - or risk - over time. We emphasize the importance of this assumption and the misleading conclusions that can be inferred if it is violated; this is particularly essential in the presence of long follow-ups. METHODS: We illustrate our discussion by analyzing prognostic factors of metastases in 979 women treated for breast cancer with surgery. Age, tumour size and grade, lymph node involvement, peritumoral vascular invasion (PVI), status of hormone receptors (HRec), Her2, and Mib1 were considered. RESULTS: Median follow-up was 14 years; 264 women developed metastases. The conventional Cox model suggested that all factors but HRec, Her2, and Mib1 status were strong prognostic factors of metastases. Additional tests indicated that the PH assumption was not satisfied for some variables of the model. Tumour grade had a significant time-varying effect, but although its effect diminished over time, it remained strong. Interestingly, while the conventional Cox model did not show any significant effect of the HRec status, tests provided strong evidence that this variable had a non-constant effect over time. Negative HRec status increased the risk of metastases early but became protective thereafter. This reversal of effect may explain non-significant hazard ratios provided by previous conventional Cox analyses in studies with long follow-ups. CONCLUSIONS: Investigating time-varying effects should be an integral part of Cox survival analyses. Detecting and accounting for time-varying effects provide insights on some specific time patterns, and on valuable biological information that could be missed otherwise

    Garbage in, garbage out: how reliable training data improved a virtual screening approach against SARS-CoV-2 MPro

    Get PDF
    Introduction: The identification of chemical compounds that interfere with SARS-CoV-2 replication continues to be a priority in several academic and pharmaceutical laboratories. Computational tools and approaches have the power to integrate, process and analyze multiple data in a short time. However, these initiatives may yield unrealistic results if the applied models are not inferred from reliable data and the resulting predictions are not confirmed by experimental evidence.Methods: We undertook a drug discovery campaign against the essential major protease (MPro) from SARS-CoV-2, which relied on an in silico search strategy –performed in a large and diverse chemolibrary– complemented by experimental validation. The computational method comprises a recently reported ligand-based approach developed upon refinement/learning cycles, and structure-based approximations. Search models were applied to both retrospective (in silico) and prospective (experimentally confirmed) screening.Results: The first generation of ligand-based models were fed by data, which to a great extent, had not been published in peer-reviewed articles. The first screening campaign performed with 188 compounds (46 in silico hits and 100 analogues, and 40 unrelated compounds: flavonols and pyrazoles) yielded three hits against MPro (IC50 ≀ 25 ÎŒM): two analogues of in silico hits (one glycoside and one benzo-thiazol) and one flavonol. A second generation of ligand-based models was developed based on this negative information and newly published peer-reviewed data for MPro inhibitors. This led to 43 new hit candidates belonging to different chemical families. From 45 compounds (28 in silico hits and 17 related analogues) tested in the second screening campaign, eight inhibited MPro with IC50 = 0.12–20 ÎŒM and five of them also impaired the proliferation of SARS-CoV-2 in Vero cells (EC50 7–45 ÎŒM).Discussion: Our study provides an example of a virtuous loop between computational and experimental approaches applied to target-focused drug discovery against a major and global pathogen, reaffirming the well-known “garbage in, garbage out” machine learning principle
    • 

    corecore