146 research outputs found

    Can analyses of electronic patient records be independently and externally validated? The effect of statins on the mortality of patients with ischaemic heart disease: a cohort study with nested case-control analysis

    Get PDF
    Objective To conduct a fully independent and external validation of a research study based on one electronic health record database, using a different electronic database sampling the same population. Design Using the Clinical Practice Research Datalink (CPRD), we replicated a published investigation into the effects of statins in patients with ischaemic heart disease (IHD) by a different research team using QResearch. We replicated the original methods and analysed all-cause mortality using: (1) a cohort analysis and (2) a case-control analysis nested within the full cohort. Setting Electronic health record databases containing longitudinal patient consultation data from large numbers of general practices distributed throughout the UK. Participants CPRD data for 34 925 patients with IHD from 224 general practices, compared to previously published results from QResearch for 13 029 patients from 89 general practices. The study period was from January 1996 to December 2003. Results We successfully replicated the methods of the original study very closely. In a cohort analysis, risk of death was lower by 55% for patients on statins, compared with 53% for QResearch (adjusted HR 0.45, 95% CI 0.40 to 0.50; vs 0.47, 95% CI 0.41 to 0.53). In case-control analyses, patients on statins had a 31% lower odds of death, compared with 39% for QResearch (adjusted OR 0.69, 95% CI 0.63 to 0.75; vs OR 0.61, 95% CI 0.52 to 0.72). Results were also close for individual statins. Conclusions Database differences in population characteristics and in data definitions, recording, quality and completeness had a minimal impact on key statistical outputs. The results uphold the validity of research using CPRD and QResearch by providing independent evidence that both datasets produce very similar estimates of treatment effect, leading to the same clinical and policy decisions. Together with other non-independent replication studies, there is a nascent body of evidence for wider validity

    Review of early hospitalisation after percutaneous coronary intervention

    Get PDF
    Background: Percutaneous coronary intervention (PCI) is the most common modality of revascularization in patients with coronary artery disease. Understanding the readmission rates and reasons for readmission after PCI is important because readmissions are a quality of care indicator, in addition to being a burden to patients and healthcare services. Methods: A literature review was performed. Relevant studies are described by narrative synthesis with the use of tables to summarize study results. Results: Data suggests that 30-day readmissions are not uncommon. The rate of readmission after PCI is highly influenced by the cohort and the healthcare system studied, with 30-day readmission rates reported to be between 4.7‐% and 15.6%. Studies consistently report that a majority of readmissions within 30 days are due to a cardiac-related disorders or complication-related disorders. Female sex, peripheral vascular disease, diabetes mellitus, renal failure and non-elective PCI are predictive of readmission. Studies also suggest that there is greater risk of mortality among patients who are readmitted compared to those who are not readmitted. Conclusion: Readmission after PCI is common and its rate is highly influenced by the type of cohort studied. There is clear evidence that majority of readmissions within 30 days are cardiac related. While there are many predictors of readmission following PCI, it is not known whether targeting patients with modifiable predictors could prevent or reduce the rates of readmission

    Transformative Machine Learning

    Get PDF
    The key to success in machine learning (ML) is the use of effective data representations. Traditionally, data representations were hand-crafted. Recently it has been demonstrated that, given sufficient data, deep neural networks can learn effective implicit representations from simple input representations. However, for most scientific problems, the use of deep learning is not appropriate as the amount of available data is limited, and/or the output models must be explainable. Nevertheless, many scientific problems do have significant amounts of data available on related tasks, which makes them amenable to multi-task learning, i.e. learning many related problems simultaneously. Here we propose a novel and general representation learning approach for multi-task learning that works successfully with small amounts of data. The fundamental new idea is to transform an input intrinsic data representation (i.e., handcrafted features), to an extrinsic representation based on what a pre-trained set of models predict about the examples. This transformation has the dual advantages of producing significantly more accurate predictions, and providing explainable models. To demonstrate the utility of this transformative learning approach, we have applied it to three real-world scientific problems: drug-design (quantitative structure activity relationship learning), predicting human gene expression (across different tissue types and drug treatments), and meta-learning for machine learning (predicting which machine learning methods work best for a given problem). In all three problems, transformative machine learning significantly outperforms the best intrinsic representation

    A Voting Ensemble Method to Assist the Diagnosis of Prostate Cancer Using Multiparametric MRI

    Get PDF
    © 2020, Springer Nature Switzerland AG. Prostate cancer is the second most commonly occurring cancer in men. Diagnosis through Magnetic Resonance Imaging (MRI) is limited, yet current practice holds a relatively low specificity. This paper extends a previous SPIE ProstateX challenge study in three ways (1) to include healthy tissue analysis, creating a solution suitable for clinical practice, which has been requested and validated by collaborating clinicians; (2) by using a voting ensemble method to assist prostate cancer diagnosis through a supervised SVM approach; and (3) using the unsupervised GTM to provide interpretability to understand the supervised SVM classification results. Pairwise classifiers of clinically significant lesion, non-significant lesion, and healthy tissue, were developed. Results showed that when combining multiparametric MRI and patient level metadata, classification of significant lesions against healthy tissue attained an AUC of 0.869 (10-fold cross-validation)

    Automatic relevance source determination in human brain tumors using Bayesian NMF.

    Get PDF
    The clinical management of brain tumors is very sensitive; thus, their non-invasive characterization is often preferred. Non-negative Matrix Factorization techniques have been successfully applied in the context of neuro-oncology to extract the underlying source signals that explain different tissue tumor types, for which knowing the number of sources to calculate was always required. In the current study we estimate the number of relevant sources for a set of discrimination problems involving brain tumors and normal brain. For this, we propose to start by calculating a high number of sources using Bayesian NMF and automatically discarding the irrelevant ones during the iterative process of matrices decomposition, hence obtaining a reduced range of interpretable solutions. The real data used in this study come from a widely tested human brain tumor database. Simulated data that resembled the real data was also generated to validate the hypothesis against ground truth. The results obtained suggest that the proposed approach is able to provide a small range of meaningful solutions to the problem of source extraction in human brain tumors

    The INTERPRET Decision-Support System version 3.0 for evaluation of Magnetic Resonance Spectroscopy data from human brain tumours and other abnormal brain masses.

    Get PDF
    Background Proton Magnetic Resonance (MR) Spectroscopy (MRS) is a widely available technique for those clinical centres equipped with MR scanners. Unlike the rest of MR-based techniques, MRS yields not images but spectra of metabolites in the tissues. In pathological situations, the MRS profile changes and this has been particularly described for brain tumours. However, radiologists are frequently not familiar to the interpretation of MRS data and for this reason, the usefulness of decision-support systems (DSS) in MRS data analysis has been explored. Results This work presents the INTERPRET DSS version 3.0, analysing the improvements made from its first release in 2002. Version 3.0 is aimed to be a program that 1st, can be easily used with any new case from any MR scanner manufacturer and 2nd, improves the initial analysis capabilities of the first version. The main improvements are an embedded database, user accounts, more diagnostic discrimination capabilities and the possibility to analyse data acquired under additional data acquisition conditions. Other improvements include a customisable graphical user interface (GUI). Most diagnostic problems included have been addressed through a pattern-recognition based approach, in which classifiers based on linear discriminant analysis (LDA) were trained and tested. Conclusions The INTERPRET DSS 3.0 allows radiologists, medical physicists, biochemists or, generally speaking, any person with a minimum knowledge of what an MR spectrum is, to enter their own SV raw data, acquired at 1.5 T, and to analyse them. The system is expected to help in the categorisation of MR Spectra from abnormal brain masses

    Impact of co-morbid burden on mortality in patients with coronary heart disease, heart failure, and cerebrovascular accident: a systematic review and meta-analysis.

    Get PDF
    Aims: We sought to investigate the prognostic impact of co-morbid burden as defined by the Charlson Co-morbidity Index (CCI) in patients with a range of prevalent cardiovascular diseases. Methods and results: We searched MEDLINE and EMBASE to identify studies that evaluated the impact of CCI on mortality in patients with cardiovascular disease. A random-effects meta-analysis was undertaken to evaluate the impact of CCI on mortality in patients with coronary heart disease (CHD), heart failure (HF), and cerebrovascular accident (CVA). A total of 11 studies of acute coronary syndrome (ACS), 2 stable coronary disease, 5 percutaneous coronary intervention (PCI), 13 HF, and 4 CVA met the inclusion criteria. An increase in CCI score per point was significantly associated with a greater risk of mortality in patients with ACS [pooled relative risk ratio (RR) 1.33; 95% CI 1.15-1.54], PCI (RR 1.21; 95% CI 1.12-1.31), stable coronary artery disease (RR 1.38; 95% CI 1.29-1.48), and HF (RR 1.21; 95% CI 1.13-1.29), but not CVA. A CCI score of >2 significantly increased the risk of mortality in ACS (RR 2.52; 95% CI 1.58-4.04), PCI (RR 3.36; 95% CI 2.14-5.29), HF (RR 1.76; 95% CI 1.65-1.87), and CVA (RR 3.80; 95% CI 1.20-12.01). Conclusion: Increasing co-morbid burden as defined by CCI is associated with a significant increase in risk of mortality in patients with underlying CHD, HF, and CVA. CCI provides a simple way of predicting adverse outcomes in patients with cardiovascular disease and should be incorporated into decision-making processes when counselling patients

    Externally validated models for first diagnosis and risk of progression of knee osteoarthritis.

    Get PDF
    ObjectiveWe develop and externally validate two models for use with radiological knee osteoarthritis. They consist of a diagnostic model for KOA and a prognostic model of time to onset of KOA. Model development and optimisation used data from the Osteoarthritis initiative (OAI) and external validation for both models was by application to data from the Multicenter Osteoarthritis Study (MOST).Materials and methodsThe diagnostic model at first presentation comprises subjects in the OAI with and without KOA (n = 2006), modelling with multivariate logistic regression. The prognostic sample involves 5-year follow-up of subjects presenting without clinical KOA (n = 1155), with modelling with Cox regression. In both instances the models used training data sets of n = 1353 and 1002 subjects and optimisation used test data sets of n = 1354 and 1003. The external validation data sets for the diagnostic and prognostic models comprised n = 2006 and n = 1155 subjects respectively.ResultsThe classification performance of the diagnostic model on the test data has an AUC of 0.748 (0.721-0.774) and 0.670 (0.631-0.708) in external validation. The survival model has concordance scores for the OAI test set of 0.74 (0.7325-0.7439) and in external validation 0.72 (0.7190-0.7373). The survival approach stratified the population into two risk cohorts. The separation between the cohorts remains when the model is applied to the validation data.DiscussionThe models produced are interpretable with app interfaces that implement nomograms. The apps may be used for stratification and for patient education over the impact of modifiable risk factors. The externally validated results, by application to data from a substantial prospective observational study, show the robustness of models for likelihood of presenting with KOA at an initial assessment based on risk factors identified by the OAI protocol and stratification of risk for developing KOA in the next five years.ConclusionModelling clinical KOA from OAI data validates well for the MOST data set. Both risk models identified key factors for differentiation of the target population from commonly available variables. With this analysis there is potential to improve clinical management of patients

    Machine Learning applications for Cataclysmic Variable discovery in the ZTF alert stream

    Get PDF
    Cataclysmic variables (CV) encompass a diverse array of accreting white dwarf binary systems. Each class of CV represents a snapshot along an evolutionary journey, one with the potential to trigger a type Ia supernova event. The study of CVs offers valuable insights into binary evolution and accretion physics, with the rarest examples potentially providing the deepest insights. However, the escalating number of detected transients, coupled with our limited capacity to investigate them all, poses challenges in identifying such rarities. Machine Learning (ML) plays a pivotal role in addressing this issue by facilitating the categorisation of each detected transient into its respective transient class. Leveraging these techniques, we have developed a two-stage pipeline tailored to the ZTF transient alert stream. The first stage is an alerts filter aimed at removing non-CVs, while the latter is an ML classifier produced using XGBoost, achieving a macro average AUC score of 0.92 for distinguishing between CV classes. By utilising the Generative Topographic Mapping algorithm with classifier posterior probabilities as input, we obtain representations indicating that CV evolutionary factors play a role in classifier performance, while the associated feature maps present a potent tool for identifying the features deemed most relevant for distinguishing between classes. Implementation of the pipeline in June 2023 yielded 51 intriguing candidates that are yet to be reported as CVs or classified with further granularity. Our classifier represents a significant step in the discovery and classification of different CV classes, a domain of research still in its infancy
    corecore