535 research outputs found

    A New Scalable, Portable, and Memory-Efficient Predictive Analytics Framework for Predicting Time-to-Event Outcomes in Healthcare

    Get PDF
    Time-to-event outcomes are prevalent in medical research. To handle these outcomes, as well as censored observations, statistical and survival regression methods are widely used based on the assumptions of linear association; however, clinicopathological features often exhibit nonlinear correlations. Machine learning (ML) algorithms have been recently adapted to effectively handle nonlinear correlations. One drawback of ML models is that they can model idiosyncratic features of a training dataset. Due to this overlearning, ML models perform well on the training data but are not so striking on test data. The features that we choose indirectly influence the performance of ML prediction models. With the expansion of big data in biomedical informatics, appropriate feature engineering and feature selection are vital to ML success. Also, an ensemble learning algorithm helps decrease bias and variance by combining the predictions of multiple models. In this study, we newly constructed a scalable, portable, and memory-efficient predictive analytics framework, fitting four components (feature engineering, survival analysis, feature selection, and ensemble learning) together. Our framework first employs feature engineering techniques, such as binarization, discretization, transformation, and normalization on raw dataset. The normalized feature set was applied to the Cox survival regression that produces highly correlated features relevant to the outcome.The resultant feature set was deployed to “eXtreme gradient boosting ensemble learning” (XGBoost) and Recursive Feature Elimination algorithms. XGBoost uses a gradient boosting decision tree algorithm in which new models are created sequentially that predict the residuals of prior models, which are then added together to make the final prediction. In our experiments, we analyzed a cohort of cardiac surgery patients drawn from a multi-hospital academic health system. The model evaluated 72 perioperative variables that impact an event of readmission within 30 days of discharge, derived 48 significant features, and demonstrated optimum predictive ability with feature sets ranging from 16 to 24. The area under the receiver operating characteristics observed for the feature set of 16 were 0.8816, and 0.9307 at the 35th, and 151st iteration respectively. Our model showed improved performance compared to state-of-the-art models and could be more useful for decision support in clinical settings

    Analyzing Patients’ EHR: Predicting and Explaining Admission Consequences for COPD and Liver Disease Patients

    Get PDF
    This study analyzed the admission outcomes in chronic patients (with COPD, and Liver disease) to demonstrate the feasibility of applying prediction methods on EHR records while incorporating an explainable AI technique. We predicted three target variables: 30-day readmission, Medium&Long Length of Stay and Single-day admission and analyzed the features using an explainable AI technique, the SHapley Additive exPlanations (SHAP). The results show that Readmission had higher prediction scores than all other dependent variables. Some features affected all target variables with either positive or negative influence including: Age, Charlson comorbidity index, Day-Shift, Gender, using EHR screens and Insurance cover level. These findings thus point to the value of using Machine-Learning combined with an explainable AI method to understand and assess the risks factors. The assessment of the potential factors leading to multiple complications can bolster prevention-oriented medical decisions to groups of patients but can also be tailored to the patient level

    RCHE Semi-Annual Report June 2013

    Get PDF

    Dynamic Prediction for Alternating Recurrent Events Using a Semiparametric Joint Frailty Model

    Get PDF
    Alternating recurrent events data arise commonly in health research; examples include hospital admissions and discharges of diabetes patients; exacerbations and remissions of chronic bronchitis; and quitting and restarting smoking. Recent work has involved formulating and estimating joint models for the recurrent event times considering non-negligible event durations. However, prediction models for transition between recurrent events are lacking. We consider the development and evaluation of methods for predicting future events within these models. Specifically, we propose a tool for dynamically predicting transition between alternating recurrent events in real time. Under a flexible joint frailty model, we derive the predictive probability of a transition from one event type to the other within a pre-specified time period. To circumvent numerical integration in calculating the predictive probability, we obtain the approximate transition probability by a Taylor expansion. Simulation results demonstrate that our tool provides better prediction performance in discrimination, as measured by the area under the ROC curve (AUC) and sensitivity, than prediction approaches that rely on standard binary regression models. Also, simulation shows that prediction results from approximate transition probability are as close as results from the exact predictive probability. We illustrate predictions in analyses of relapses of chronic bronchitis exacerbation from a pharmaceutical trial and hospital readmissions in patients with diabetes from Medicaid claims data. The final part of this dissertation (Chapter 6) compares predictive performance between logistic regression and random forests for 30-day readmission using longitudinal claims data. Several studies have compared these and other prediction models using longitudinal electronic health records or claims data. Because most of them applied logistic regression to the longitudinal observations, ignoring the lack of independence within subjects, or claims data consisting of independent observations, a correct comparison of the models under longitudinal data remains obscure. Moreover, those studies did not compare the out-of-sample performance. We address these issues and compare the prediction performance of the models using longitudinal claims data. We implement simulations by randomly choosing a record from each patient\u27s multiple records in the training set, fitting the two models, applying the models to the training, test, and external sets, and obtaining AUC and sensitivity for each. We observe that although random forests generally gives better predictions on the training set, logistic regression performs better on test and external sets. In an empirical study, we apply the prediction methods to Medicaid claims data covering inpatient admissions of patients with heart failure

    Ensemble Risk Model of Emergency Admissions (ERMER)

    Get PDF
    Introduction About half of hospital readmissions can be avoided with preventive interventions. Developing decision support tools for identification of patients’ emergency readmission risk is an important area of research. Because, it remains unclear how to design features and develop predictive models that can adjust continuously to a fast-changing healthcare system and population characteristics. The objective of this study was to develop a generic ensemble Bayesian risk model of emergency readmission. Methods We produced a decision support tool that predicts risk of emergency readmission using England's Hospital Episode Statistics inpatient database. Firstly, we used a framework to develop an optimal set of features. Then, a combination of Bayes Point Machine (BPM) models for different cohorts was considered to create an optimised ensemble model, which is stronger than the individual generative and non-linear classifications. The developed Ensemble Risk Model of Emergency Admissions (ERMER) was trained and tested using three time-frames: 1999-2004, 2000-05 and 2004-09, each of which includes about 20% of patients in England during the trigger year. Results Comparisons are made for different time-frames, sub-populations, risk cut-offs, risk bands and top risk segments. The precision was 71.6% to 73.9%, the specificity was 88.3% to 91.7% and the sensitivity was 42.1% to 49.2% across different time-frames. Moreover, the Area Under the Curve was 75.9% to 77.1%. Conclusions The decision support tool performed considerably better than the previous modelling approaches, and it was robust and stable with high precision. Moreover, the framework and the Bayesian model allow the model to continuously adjust it to new significant features, different population characteristics and changes in the system

    Predicting risk of hospital readmission for comorbidity patients through a novel deep learning framework

    Get PDF
    Hospital readmission is widely recognized as indicator of inpatient quality of care which has significant impact on healthcare cost. Thus, early recognition of readmission risk has been of growing interest in various hospitals. Additionally, there has been growing attention to provide better care to patients with more complications, whose care would impact the quality of care in multiple directions. To this regard, this research specifically targets comorbidity patients i.e., the patients with chronic disease. This research proposes a novel deep learning- framework termed SDAE-GAN. The presented approach consists of three phases. Firstly, various groups of variables from heterogeneous sources are collated. These variables mainly include demographic, socioeconomic, some statistics about patient’s frequent admissions and their diagnosis codes. Then, more processing applies dealing missing values, digitization and data balancing. Afterwards, stacked denoising auto-encoders function to learn underlying representation; and technically to forms a latent space. The latent variables then are used by a Generative Adversarial Neural Networks to evaluate the risk of 30- day readmission. The model is fine-tuned and being compared with state-of-the-arts. Experimental results exhibit competitive performance with higher sensitivity

    Scoping Meta-Review of Methods Used to Assess Artificial Intelligence-Based Medical Devices for Heart Failure

    Get PDF
    Artificial intelligence and machine learning (AI/ML) are playing increasingly important roles, permeating the field of medical devices (MDs). This rapid progress has not yet been matched by the Health Technology Assessment (HTA) process, which still needs to define a common methodology for assessing AI/ML-based MDs. To collect existing evidence from the literature about the methods used to assess AI-based MDs, with a specific focus on those used for the management of heart failure (HF), the International Federation of Medical and Biological Engineering (IFMBE) conducted a scoping meta-review. This manuscript presents the results of this search, which covered the period from January 1974 to October 2022. After careful independent screening, 21 reviews, mainly conducted in North America and Europe, were retained and included. Among the findings were that deep learning is the most commonly utilised method and that electronic health records and registries are among the most prevalent sources of data for AI/ML algorithms. Out of the 21 included reviews, 19 focused on risk prediction and/or the early diagnosis of HF. Furthermore, 10 reviews provided evidence of the impact on the incidence/progression of HF, and 13 on the length of stay. From an HTA perspective, the main areas requiring improvement are the quality assessment of studies on AI/ML (included in 11 out of 21 reviews) and their data sources, as well as the definition of the criteria used to assess the selection of the most appropriate AI/ML algorithm. © 2023 by the authors
    • 

    corecore