3,377 research outputs found

    Machine Learning Time-to-Event Mortality Prediction in MIMIC-IV Critical Care Database

    Get PDF
    The rise in publicly available healthcare databases, such as MIMIC and the eICU, now make it possible to revolutionize medical care when paired with modern machine learning techniques. The MIMIC-IV critical care database allows us to explore these techniques in the ICU setting using data from thousands of patients. One area that can be improved upon in the medical domain is prediction of events in the ICU setting, such as whether a patient will have a heart attack during their stay. Through improved prediction of events, hospitals can be more efficient and better allocate resources to patients who need it most, saving both lives and costs. In the ICU setting, there has been previous work for prediction of events via machine learning classification models. However, we believe time-to-event models may offer more accuracy and interpretability than these classification models. We also believe that current, popular time-to-event models are limited in their scope, either not being able to deal with dynamic data, being too slow to use in real time, or having to make assumptions about the underlying structure of the data. Some models also require restructuring of the data into specific formats which leads to information loss. These kinds of restrictions are not desirable in the ICU, where measurements come in frequently at irregular intervals and requires fast prediction of events. It follows, then, that we need dynamic, lightweight, time-to-event models for prediction of events that do not make assumptions about the data’s structure. In this paper, we use BoXHED, a lightweight, dynamic, boosting, time-to-event model and compare it to other time-varying models in the ICU. To evaluate the different models’ performances, we used time series data from the MIMIC-IV database by refactoring code used previously for MIMIC-III preprocessing by Harutyunyan et al. We then compared the different models’ accuracy in predicting mortality in the ICU, both as new data measurements became available and using measurements within the first 48 hours of the patients’ stays. We then evaluated the models based on an approach inspired by TREWScore where patient risk scores were compared to given thresholds to obtain each models’ AUC-ROC scores

    Continuous and automatic mortality risk prediction using vital signs in the intensive care unit: a hybrid neural network approach

    Get PDF
    Mortality risk prediction can greatly improve the utilization of resources in intensive care units (ICUs). Existing schemes in ICUs today require laborious manual input of many complex parameters. In this work, we present a scheme that uses variations in vital signs over a 24-h period to make mortality risk assessments for 3-day, 7-day, and 14-day windows. We develop a hybrid neural network model that combines convolutional (CNN) layers with bidirectional long short-term memory (BiLSTM) to predict mortality from statistics describing the variation of heart rate, blood pressure, respiratory rate, blood oxygen levels, and temperature. Our scheme performs strongly compared to state-of-the-art schemes in the literature for mortality prediction, with our highest-performing model achieving an area under the receiver-operator curve of 0.884. We conclude that the use of a hybrid CNN-BiLSTM network is highly effective in determining mortality risk for the 3, 7, and 14 day windows from vital signs. As vital signs are routinely recorded, in many cases automatically, our scheme could be implemented such that highly accurate mortality risk could be predicted continuously and automatically, reducing the burden on healthcare providers and improving patient outcomes

    Federated Learning for Mortality Prediction in Intensive Care Units

    Get PDF
    Federated learning is a method to train a machine learning model on multiple remote datasets without the need to gather the data from the remote sites to a central location. In healthcare, gathering the data from different hospitals into a central location can be a difficult and time-consuming task, due to privacy concerns and regulations regarding the use of sensitive data, making federated learning an attractive alternative to more traditional methods. This thesis adapted an existing federated gradient boosting model and developed a new federated random forest model and applied them to mortality prediction in intensive care units. The results were then compared to the centralized counterparts of the models. The results showed that while the federated models did not perform as well as the centralized models on a similar sized dataset, the federated random forest model can achieve superior performance when trained on multiple hospitals' data compared to centralized models trained on a single hospital. In scenarios where the centralized models had data from multiple hospitals the federated models could not perform as well as the centralized models. It was also found that the performance of the centralized models could not be improved with further federated training. In addition to practical advantages such as possibility of parallel or asynchronous training without modifications to the algorithm, the federated random forest performed better in all scenarios compared to the federated gradient boosting. The performance of the federated random forest was also found to be more consistent over different scenarios than the performance of federated gradient boosting, which was highly dependent on factors such as the order with the hospitals were traversed

    Toward More Predictive Models by Leveraging Multimodal Data

    Get PDF
    Data is often composed of structured and unstructured data. Both forms of data have information that can be exploited by machine learning models to increase their prediction performance on a task. However, integrating the features from both these data forms is a hard, complicated task. This is all the more true for models which operate on time-constraints. Time-constrained models are machine learning models that work on input where time causality has to be maintained such as predicting something in the future based on past data. Most previous work does not have a dedicated pipeline that is generalizable to different tasks and domains, especially under time-constraints. In this work, we present a systematic, domain-agnostic pipeline for integrating features from structured and unstructured data while maintaining time causality for building models. We focus on the healthcare and consumer market domain and perform experiments, preprocess data, and build models to demonstrate the generalizability of the pipeline. More specifically, we focus on the task of identifying patients who are at risk of an imminent ICU admission. We use our pipeline to solve this task and show how augmenting unstructured data with structured data improves model performance. We found that by combining structured and unstructured data we can get a performance improvement of up to 8.5

    Learning Credible Models

    Full text link
    In many settings, it is important that a model be capable of providing reasons for its predictions (i.e., the model must be interpretable). However, the model's reasoning may not conform with well-established knowledge. In such cases, while interpretable, the model lacks \textit{credibility}. In this work, we formally define credibility in the linear setting and focus on techniques for learning models that are both accurate and credible. In particular, we propose a regularization penalty, expert yielded estimates (EYE), that incorporates expert knowledge about well-known relationships among covariates and the outcome of interest. We give both theoretical and empirical results comparing our proposed method to several other regularization techniques. Across a range of settings, experiments on both synthetic and real data show that models learned using the EYE penalty are significantly more credible than those learned using other penalties. Applied to a large-scale patient risk stratification task, our proposed technique results in a model whose top features overlap significantly with known clinical risk factors, while still achieving good predictive performance

    Predictive analytics framework for electronic health records with machine learning advancements : optimising hospital resources utilisation with predictive and epidemiological models

    Get PDF
    The primary aim of this thesis was to investigate the feasibility and robustness of predictive machine-learning models in the context of improving hospital resources’ utilisation with data- driven approaches and predicting hospitalisation with hospital quality assessment metrics such as length of stay. The length of stay predictions includes the validity of the proposed methodological predictive framework on each hospital’s electronic health records data source. In this thesis, we relied on electronic health records (EHRs) to drive a data-driven predictive inpatient length of stay (LOS) research framework that suits the most demanding hospital facilities for hospital resources’ utilisation context. The thesis focused on the viability of the methodological predictive length of stay approaches on dynamic and demanding healthcare facilities and hospital settings such as the intensive care units and the emergency departments. While the hospital length of stay predictions are (internal) healthcare inpatients outcomes assessment at the time of admission to discharge, the thesis also considered (external) factors outside hospital control, such as forecasting future hospitalisations from the spread of infectious communicable disease during pandemics. The internal and external splits are the thesis’ main contributions. Therefore, the thesis evaluated the public health measures during events of uncertainty (e.g. pandemics) and measured the effect of non-pharmaceutical intervention during outbreaks on future hospitalised cases. This approach is the first contribution in the literature to examine the epidemiological curves’ effect using simulation models to project the future hospitalisations on their strong potential to impact hospital beds’ availability and stress hospital workflow and workers, to the best of our knowledge. The main research commonalities between chapters are the usefulness of ensembles learning models in the context of LOS for hospital resources utilisation. The ensembles learning models anticipate better predictive performance by combining several base models to produce an optimal predictive model. These predictive models explored the internal LOS for various chronic and acute conditions using data-driven approaches to determine the most accurate and powerful predicted outcomes. This eventually helps to achieve desired outcomes for hospital professionals who are working in hospital settings

    A dynamic risk score for early prediction of cardiogenic shock using machine learning

    Full text link
    Myocardial infarction and heart failure are major cardiovascular diseases that affect millions of people in the US. The morbidity and mortality are highest among patients who develop cardiogenic shock. Early recognition of cardiogenic shock is critical. Prompt implementation of treatment measures can prevent the deleterious spiral of ischemia, low blood pressure, and reduced cardiac output due to cardiogenic shock. However, early identification of cardiogenic shock has been challenging due to human providers' inability to process the enormous amount of data in the cardiac intensive care unit (ICU) and lack of an effective risk stratification tool. We developed a deep learning-based risk stratification tool, called CShock, for patients admitted into the cardiac ICU with acute decompensated heart failure and/or myocardial infarction to predict onset of cardiogenic shock. To develop and validate CShock, we annotated cardiac ICU datasets with physician adjudicated outcomes. CShock achieved an area under the receiver operator characteristic curve (AUROC) of 0.820, which substantially outperformed CardShock (AUROC 0.519), a well-established risk score for cardiogenic shock prognosis. CShock was externally validated in an independent patient cohort and achieved an AUROC of 0.800, demonstrating its generalizability in other cardiac ICUs

    Integrated Machine Learning and Optimization Frameworks with Applications in Operations Management

    Full text link
    Incorporation of contextual inference in the optimality analysis of operational problems is a canonical characteristic of data-informed decision making that requires interdisciplinary research. In an attempt to achieve individualization in operations management, we design rigorous and yet practical mechanisms that boost efficiency, restrain uncertainty and elevate real-time decision making through integration of ideas from machine learning and operations research literature. In our first study, we investigate the decision of whether to admit a patient to a critical care unit which is a crucial operational problem that has significant influence on both hospital performance and patient outcomes. Hospitals currently lack a methodology to selectively admit patients to these units in a way that patient’s individual health metrics can be incorporated while considering the hospital’s operational constraints. We model the problem as a complex loss queueing network with a stochastic model of how long risk-stratified patients spend time in particular units and how they transition between units. A data-driven optimization methodology then approximates an optimal admission control policy for the network of units. While enforcing low levels of patient blocking, we optimize a monotonic dual-threshold admission policy. Our methodology captures utilization and accessibility in a network model of care pathways while supporting the personalized allocation of scarce care resources to the neediest patients. The interesting benefits of admission thresholds that vary by day of week are also examined. In the second study, we analyze the efficiency of surgical unit operations in the era of big data. The accuracy of surgical case duration predictions is a crucial element in hospital operational performance. We propose a comprehensive methodology that incorporates both structured and unstructured data to generate individualized predictions regarding the overall distribution of surgery durations. Consequently, we investigate methods to incorporate such individualized predictions into operational decision-making. We introduce novel prescriptive models to address optimization under uncertainty in the fundamental surgery appointment scheduling problem by utilizing the multi-dimensional data features available prior to the surgery. Electronic medical records systems provide detailed patient features that enable the prediction of individualized case time distributions; however, existing approaches in this context usually employ only limited, aggregate information, and do not take advantages of these detailed features. We show how the quantile regression forest, can be integrated into three common optimization formulations that capture the stochasticity in addressing this problem, including stochastic optimization, robust optimization and distributionally robust optimization. In the last part of this dissertation, we provide the first study on online learning problems under stochastic constraints that are "soft", i.e., need to be satisfied with high likelihood. Under a Bayesian framework, we propose and analyze a scheme that provides statistical feasibility guarantees throughout the learning horizon, by using posterior Monte Carlo samples to form sampled constraints that generalize the scenario generation approach commonly used in chance-constrained programming. We demonstrate how our scheme can be integrated into Thompson sampling and illustrate it with an application in online advertisement.PHDIndustrial & Operations EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttps://deepblue.lib.umich.edu/bitstream/2027.42/145936/1/meisami_1.pd
    • …
    corecore