77 research outputs found

    Deep Learning in Cardiology

    Full text link
    The medical field is creating large amount of data that physicians are unable to decipher and use efficiently. Moreover, rule-based expert systems are inefficient in solving complicated medical tasks or for creating insights using big data. Deep learning has emerged as a more accurate and effective technology in a wide range of medical problems such as diagnosis, prediction and intervention. Deep learning is a representation learning method that consists of layers that transform the data non-linearly, thus, revealing hierarchical relationships and structures. In this review we survey deep learning application papers that use structured data, signal and imaging modalities from cardiology. We discuss the advantages and limitations of applying deep learning in cardiology that also apply in medicine in general, while proposing certain directions as the most viable for clinical use.Comment: 27 pages, 2 figures, 10 table

    Deep Risk Prediction and Embedding of Patient Data: Application to Acute Gastrointestinal Bleeding

    Get PDF
    Acute gastrointestinal bleeding is a common and costly condition, accounting for over 2.2 million hospital days and 19.2 billion dollars of medical charges annually. Risk stratification is a critical part of initial assessment of patients with acute gastrointestinal bleeding. Although all national and international guidelines recommend the use of risk-assessment scoring systems, they are not commonly used in practice, have sub-optimal performance, may be applied incorrectly, and are not easily updated. With the advent of widespread electronic health record adoption, longitudinal clinical data captured during the clinical encounter is now available. However, this data is often noisy, sparse, and heterogeneous. Unsupervised machine learning algorithms may be able to identify structure within electronic health record data while accounting for key issues with the data generation process: measurements missing-not-at-random and information captured in unstructured clinical note text. Deep learning tools can create electronic health record-based models that perform better than clinical risk scores for gastrointestinal bleeding and are well-suited for learning from new data. Furthermore, these models can be used to predict risk trajectories over time, leveraging the longitudinal nature of the electronic health record. The foundation of creating relevant tools is the definition of a relevant outcome measure; in acute gastrointestinal bleeding, a composite outcome of red blood cell transfusion, hemostatic intervention, and all-cause 30-day mortality is a relevant, actionable outcome that reflects the need for hospital-based intervention. However, epidemiological trends may affect the relevance and effectiveness of the outcome measure when applied across multiple settings and patient populations. Understanding the trends in practice, potential areas of disparities, and value proposition for using risk stratification in patients presenting to the Emergency Department with acute gastrointestinal bleeding is important in understanding how to best implement a robust, generalizable risk stratification tool. Key findings include a decrease in the rate of red blood cell transfusion since 2014 and disparities in access to upper endoscopy for patients with upper gastrointestinal bleeding by race/ethnicity across urban and rural hospitals. Projected accumulated savings of consistent implementation of risk stratification tools for upper gastrointestinal bleeding total approximately $1 billion 5 years after implementation. Most current risk scores were designed for use based on the location of the bleeding source: upper or lower gastrointestinal tract. However, the location of the bleeding source is not always clear at presentation. I develop and validate electronic health record based deep learning and machine learning tools for patients presenting with symptoms of acute gastrointestinal bleeding (e.g., hematemesis, melena, hematochezia), which is more relevant and useful in clinical practice. I show that they outperform leading clinical risk scores for upper and lower gastrointestinal bleeding, the Glasgow Blatchford Score and the Oakland score. While the best performing gradient boosted decision tree model has equivalent overall performance to the fully connected feedforward neural network model, at the very low risk threshold of 99% sensitivity the deep learning model identifies more very low risk patients. Using another deep learning model that can model longitudinal risk, the long-short-term memory recurrent neural network, need for transfusion of red blood cells can be predicted at every 4-hour interval in the first 24 hours of intensive care unit stay for high risk patients with acute gastrointestinal bleeding. Finally, for implementation it is important to find patients with symptoms of acute gastrointestinal bleeding in real time and characterize patients by risk using available data in the electronic health record. A decision rule-based electronic health record phenotype has equivalent performance as measured by positive predictive value compared to deep learning and natural language processing-based models, and after live implementation appears to have increased the use of the Acute Gastrointestinal Bleeding Clinical Care pathway. Patients with acute gastrointestinal bleeding but with other groups of disease concepts can be differentiated by directly mapping unstructured clinical text to a common ontology and treating the vector of concepts as signals on a knowledge graph; these patients can be differentiated using unbalanced diffusion earth mover’s distances on the graph. For electronic health record data with data missing not at random, MURAL, an unsupervised random forest-based method, handles data with missing values and generates visualizations that characterize patients with gastrointestinal bleeding. This thesis forms a basis for understanding the potential for machine learning and deep learning tools to characterize risk for patients with acute gastrointestinal bleeding. In the future, these tools may be critical in implementing integrated risk assessment to keep low risk patients out of the hospital and guide resuscitation and timely endoscopic procedures for patients at higher risk for clinical decompensation

    A CNN-LSTM for predicting mortality in the ICU

    Get PDF
    An accurate predicted mortality is crucial to healthcare as it provides an empirical risk estimate for prognostic decision making, patient stratification and hospital benchmarking. Current prediction methods in practice are severity of disease scoring systems that usually involve a fixed set of admission attributes and summarized physiological data. These systems are prone to bias and require substantial manual effort which necessitates an updated approach which can account for most shortcomings. Clinical observation notes allow for recording highly subjective data on the patient that can possibly facilitate higher discrimination. Moreover, deep learning models can automatically extract and select features without human input.This thesis investigates the potential of a combination of a deep learning model and notes for predicting mortality with a higher accuracy. A custom architecture, called CNN-LSTM, is conceptualized for mapping multiple notes compiled in a hospital stay to a mortality outcome. It employs both convolutional and recurrent layers with the former capturing semantic relationships in individual notes independently and the latter capturing temporal relationships between concurrent notes in a hospital stay. This approach is compared to three severity of disease scoring systems with a case study on the MIMIC-III dataset. Experiments are set up to assess the CNN-LSTM for predicting mortality using only the notes from the first 24, 12 and 48 hours of a patient stay. The model is trained using K-fold cross-validation with k=5 and the mortality probability calculated by the three severity scores on the held-out set is used as the baseline. It is found that the CNN-LSTM outperforms the baseline on all experiments which serves as a proof-of-concept of how notes and deep learning can better outcome prediction

    λ”₯ λ‰΄λŸ΄ λ„€νŠΈμ›Œν¬λ₯Ό ν™œμš©ν•œ μ˜ν•™ κ°œλ… 및 ν™˜μž ν‘œν˜„ ν•™μŠ΅κ³Ό 의료 λ¬Έμ œμ—μ˜ μ‘μš©

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(박사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·정보곡학뢀, 2022. 8. 정ꡐ민.λ³Έ ν•™μœ„ 논문은 μ „κ΅­λ―Ό 의료 λ³΄ν—˜λ°μ΄ν„°μΈ ν‘œλ³Έμ½”ν˜ΈνŠΈDBλ₯Ό ν™œμš©ν•˜μ—¬ λ”₯ λ‰΄λŸ΄ λ„€νŠΈμ›Œν¬ 기반의 μ˜ν•™ κ°œλ… 및 ν™˜μž ν‘œν˜„ ν•™μŠ΅ 방법과 의료 문제 ν•΄κ²° 방법을 μ œμ•ˆν•œλ‹€. λ¨Όμ € 순차적인 ν™˜μž 의료 기둝과 개인 ν”„λ‘œνŒŒμΌ 정보λ₯Ό 기반으둜 ν™˜μž ν‘œν˜„μ„ ν•™μŠ΅ν•˜κ³  ν–₯ν›„ μ§ˆλ³‘ 진단 κ°€λŠ₯성을 μ˜ˆμΈ‘ν•˜λŠ” μž¬κ·€μ‹ κ²½λ§ λͺ¨λΈμ„ μ œμ•ˆν•˜μ˜€λ‹€. μš°λ¦¬λŠ” λ‹€μ–‘ν•œ μ„±κ²©μ˜ ν™˜μž 정보λ₯Ό 효율적으둜 ν˜Όν•©ν•˜λŠ” ꡬ쑰λ₯Ό λ„μž…ν•˜μ—¬ 큰 μ„±λŠ₯ ν–₯상을 μ–»μ—ˆλ‹€. λ˜ν•œ ν™˜μžμ˜ 의료 기둝을 μ΄λ£¨λŠ” 의료 μ½”λ“œλ“€μ„ λΆ„μ‚° ν‘œν˜„μœΌλ‘œ λ‚˜νƒ€λ‚΄ μΆ”κ°€ μ„±λŠ₯ κ°œμ„ μ„ μ΄λ£¨μ—ˆλ‹€. 이λ₯Ό 톡해 의료 μ½”λ“œμ˜ λΆ„μ‚° ν‘œν˜„μ΄ μ€‘μš”ν•œ μ‹œκ°„μ  정보λ₯Ό λ‹΄κ³  μžˆμŒμ„ ν™•μΈν•˜μ˜€κ³ , μ΄μ–΄μ§€λŠ” μ—°κ΅¬μ—μ„œλŠ” μ΄λŸ¬ν•œ μ‹œκ°„μ  정보가 강화될 수 μžˆλ„λ‘ κ·Έλž˜ν”„ ꡬ쑰λ₯Ό λ„μž…ν•˜μ˜€λ‹€. μš°λ¦¬λŠ” 의료 μ½”λ“œμ˜ λΆ„μ‚° ν‘œν˜„ κ°„μ˜ μœ μ‚¬λ„μ™€ 톡계적 정보λ₯Ό 가지고 κ·Έλž˜ν”„λ₯Ό κ΅¬μΆ•ν•˜μ˜€κ³  κ·Έλž˜ν”„ λ‰΄λŸ΄ λ„€νŠΈμ›Œν¬λ₯Ό ν™œμš©, μ‹œκ°„/톡계적 정보가 κ°•ν™”λœ 의료 μ½”λ“œμ˜ ν‘œν˜„ 벑터λ₯Ό μ–»μ—ˆλ‹€. νšλ“ν•œ 의료 μ½”λ“œ 벑터λ₯Ό 톡해 μ‹œνŒ μ•½λ¬Όμ˜ 잠재적인 λΆ€μž‘μš© μ‹ ν˜Έλ₯Ό νƒμ§€ν•˜λŠ” λͺ¨λΈμ„ μ œμ•ˆν•œ κ²°κ³Ό, 기쑴의 λΆ€μž‘μš© λ°μ΄ν„°λ² μ΄μŠ€μ— μ‘΄μž¬ν•˜μ§€ μ•ŠλŠ” μ‚¬λ‘€κΉŒμ§€λ„ μ˜ˆμΈ‘ν•  수 μžˆμŒμ„ λ³΄μ˜€λ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ λΆ„λŸ‰μ— λΉ„ν•΄ μ£Όμš” 정보가 ν¬μ†Œν•˜λ‹€λŠ” 의료 기둝의 ν•œκ³„λ₯Ό κ·Ήλ³΅ν•˜κΈ° μœ„ν•΄ μ§€μ‹κ·Έλž˜ν”„λ₯Ό ν™œμš©ν•˜μ—¬ 사전 μ˜ν•™ 지식을 λ³΄κ°•ν•˜μ˜€λ‹€. μ΄λ•Œ ν™˜μžμ˜ 의료 기둝을 κ΅¬μ„±ν•˜λŠ” μ§€μ‹κ·Έλž˜ν”„μ˜ λΆ€λΆ„λ§Œμ„ μΆ”μΆœν•˜μ—¬ κ°œμΈν™”λœ μ§€μ‹κ·Έλž˜ν”„λ₯Ό λ§Œλ“€κ³  κ·Έλž˜ν”„ λ‰΄λŸ΄ λ„€νŠΈμ›Œν¬λ₯Ό 톡해 κ·Έλž˜ν”„μ˜ ν‘œν˜„ 벑터λ₯Ό νšλ“ν•˜μ˜€λ‹€. μ΅œμ’…μ μœΌλ‘œ 순차적인 의료 기둝을 ν•¨μΆ•ν•œ ν™˜μž ν‘œν˜„κ³Ό λ”λΆˆμ–΄ κ°œμΈν™”λœ μ˜ν•™ 지식을 ν•¨μΆ•ν•œ ν‘œν˜„μ„ ν•¨κ»˜ μ‚¬μš©ν•˜μ—¬ ν–₯ν›„ μ§ˆλ³‘ 및 진단 예츑 λ¬Έμ œμ— ν™œμš©ν•˜μ˜€λ‹€.This dissertation proposes a deep neural network-based medical concept and patient representation learning methods using medical claims data to solve two healthcare tasks, i.e., clinical outcome prediction and post-marketing adverse drug reaction (ADR) signal detection. First, we propose SAF-RNN, a Recurrent Neural Network (RNN)-based model that learns a deep patient representation based on the clinical sequences and patient characteristics. Our proposed model fuses different types of patient records using feature-based gating and self-attention. We demonstrate that high-level associations between two heterogeneous records are effectively extracted by our model, thus achieving state-of-the-art performances for predicting the risk probability of cardiovascular disease. Secondly, based on the observation that the distributed medical code embeddings represent temporal proximity between the medical codes, we introduce a graph structure to enhance the code embeddings with such temporal information. We construct a graph using the distributed code embeddings and the statistical information from the claims data. We then propose the Graph Neural Network(GNN)-based representation learning for post-marketing ADR detection. Our model shows competitive performances and provides valid ADR candidates. Finally, rather than using patient records alone, we utilize a knowledge graph to augment the patient representation with prior medical knowledge. Using SAF-RNN and GNN, the deep patient representation is learned from the clinical sequences and the personalized medical knowledge. It is then used to predict clinical outcomes, i.e., next diagnosis prediction and CVD risk prediction, resulting in state-of-the-art performances.1 Introduction 1 2 Background 8 2.1 Medical Concept Embedding 8 2.2 Encoding Sequential Information in Clinical Records 11 3 Deep Patient Representation with Heterogeneous Information 14 3.1 Related Work 16 3.2 Problem Statement 19 3.3 Method 20 3.3.1 RNN-based Disease Prediction Model 20 3.3.2 Self-Attentive Fusion (SAF) Encoder 23 3.4 Dataset and Experimental Setup 24 3.4.1 Dataset 24 3.4.2 Experimental Design 26 ii 3.4.3 Implementation Details 27 3.5 Experimental Results 28 3.5.1 Evaluation of CVD Prediction 28 3.5.2 Sensitivity Analysis 28 3.5.3 Ablation Studies 31 3.6 Further Investigation 32 3.6.1 Case Study: Patient-Centered Analysis 32 3.6.2 Data-Driven CVD Risk Factors 32 3.7 Conclusion 33 4 Graph-Enhanced Medical Concept Embedding 40 4.1 Related Work 42 4.2 Problem Statement 43 4.3 Method 44 4.3.1 Code Embedding Learning with Skip-gram Model 44 4.3.2 Drug-disease Graph Construction 45 4.3.3 A GNN-based Method for Learning Graph Structure 47 4.4 Dataset and Experimental Setup 49 4.4.1 Dataset 49 4.4.2 Experimental Design 50 4.4.3 Implementation Details 52 4.5 Experimental Results 53 4.5.1 Evaluation of ADR Detection 53 4.5.2 Newly-Described ADR Candidates 54 4.6 Conclusion 55 5 Knowledge-Augmented Deep Patient Representation 57 5.1 Related Work 60 5.1.1 Incorporating Prior Medical Knowledge for Clinical Outcome Prediction 60 5.1.2 Inductive KGC based on Subgraph Learning 61 5.2 Method 61 5.2.1 Extracting Personalized KG 61 5.2.2 KA-SAF: Knowledge-Augmented Self-Attentive Fusion Encoder 64 5.2.3 KGC as a Pre-training Task 68 5.2.4 Subgraph Infomax: SGI 69 5.3 Dataset and Experimental Setup 72 5.3.1 Clinical Outcome Prediction 72 5.3.2 Next Diagnosis Prediction 72 5.4 Experimental Results 73 5.4.1 Cardiovascular Disease Prediction 73 5.4.2 Next Diagnosis Prediction 73 5.4.3 KGC on SemMed KG 73 5.5 Conclusion 74 6 Conclusion 77 Abstract (In Korean) 90 Acknowlegement 92λ°•

    Methods and Techniques for Clinical Text Modeling and Analytics

    Get PDF
    Nowadays, a large portion of clinical data only exists in free text. The wide adoption of Electronic Health Records (EHRs) has enabled the increases in accessing to clinical documents, which provide challenges and opportunities for clinical Natural Language Processing (NLP) researchers. Given free-text clinical notes as input, an ideal system for clinical text understanding should have the ability to support clinical decisions. At corpus level, the system should recommend similar notes based on disease or patient types, and provide medication recommendation, or any other type of recommendations, based on patients' symptoms and other similar medical cases. At document level, it should return a list of important clinical concepts. Moreover, the system should be able to make diagnostic inferences over clinical concepts and output diagnosis. Unfortunately, current work has not systematically studied this system. This study focuses on developing and applying methods/techniques in different aspects of the system for clinical text understanding, at both corpus and document level. We deal with two major research questions: First, we explore the question of How to model the underlying relationships from clinical notes at corpus level? Documents clustering methods can group clinical notes into meaningful clusters, which can assist physicians and patients to understand medical conditions and diseases from clinical notes. We use Nonnegative Matrix Factorization (NMF) and Multi-view NMF to cluster clinical notes based on extracted medical concepts. The clustering results display latent patterns existed among clinical notes. Our method provides a feasible way to visualize a corpus of clinical documents. Based on extracted concepts, we further build a symptom-medication (Symp-Med) graph to model the Symp-Med relations in clinical notes corpus. We develop two Symp-Med matching algorithms to predict and recommend medications for patients based on their symptoms. Second, we want to solve the question of How to integrate structured knowledge with unstructured text to improve results for Clinical NLP tasks? On the one hand, the unstructured clinical text contains lots of information about medical conditions. On the other hand, structured Knowledge Bases (KBs) are frequently used for supporting clinical NLP tasks. We propose graph-regularized word embedding models to integrate knowledge from both KBs and free text. We evaluate our models on standard datasets and biomedical NLP tasks, and results showed encouraging improvements on both datasets. We further apply the graph-regularized word embedding models and present a novel approach to automatically infer the most probable diagnosis from a given clinical narrative.Ph.D., Information Studies -- Drexel University, 201

    Comprehensive Performance Analysis of Neurodegenerative disease Incidence in the Females of 60-96 year Age Group

    Get PDF
    Neurodegenerative diseases such as Alzheimer's disease and dementia are gradually becoming more prevalent chronic diseases, characterized by the decline in cognitive and behavioral symptoms. Machine learning is revolu-tionising almost all domains of our life, including the clinical system. The application of machine learning has the potential to enormously augment the reach of neurodegenerative care thus building it more proficient. Throughout the globe, there is a massive burden of Alzheimer's and demen-tia cases; which denotes an exclusive set of difficulties. This provides us with an exceptional opportunity in terms of the impending convenience of data. Harnessing this data using machine learning tools and techniques, can put scientists and physicians in the lead research position in this area. The ob-jective of this study was to develop an efficient prognostic ML model with high-performance metrics to better identify female candidate subjects at risk of having Alzheimer's disease and dementia. The study was based on two diverse datasets. The results have been discussed employing seven perfor-mance evaluation measures i.e. accuracy, precision, recall, F-measure, Re-ceiver Operating Characteristic (ROC) area, Kappa statistic, and Root Mean Squared Error (RMSE). Also, a comprehensive performance analysis has been carried out later in the study

    The Convergence of Human and Artificial Intelligence on Clinical Care - Part I

    Get PDF
    This edited book contains twelve studies, large and pilots, in five main categories: (i) adaptive imputation to increase the density of clinical data for improving downstream modeling; (ii) machine-learning-empowered diagnosis models; (iii) machine learning models for outcome prediction; (iv) innovative use of AI to improve our understanding of the public view; and (v) understanding of the attitude of providers in trusting insights from AI for complex cases. This collection is an excellent example of how technology can add value in healthcare settings and hints at some of the pressing challenges in the field. Artificial intelligence is gradually becoming a go-to technology in clinical care; therefore, it is important to work collaboratively and to shift from performance-driven outcomes to risk-sensitive model optimization, improved transparency, and better patient representation, to ensure more equitable healthcare for all

    The Application of Computer Techniques to ECG Interpretation

    Get PDF
    This book presents some of the latest available information on automated ECG analysis written by many of the leading researchers in the field. It contains a historical introduction, an outline of the latest international standards for signal processing and communications and then an exciting variety of studies on electrophysiological modelling, ECG Imaging, artificial intelligence applied to resting and ambulatory ECGs, body surface mapping, big data in ECG based prediction, enhanced reliability of patient monitoring, and atrial abnormalities on the ECG. It provides an extremely valuable contribution to the field

    Enhance Representation Learning of Clinical Narrative with Neural Networks for Clinical Predictive Modeling

    Get PDF
    Medicine is undergoing a technological revolution. Understanding human health from clinical data has major challenges from technical and practical perspectives, thus prompting methods that understand large, complex, and noisy data. These methods are particularly necessary for natural language data from clinical narratives/notes, which contain some of the richest information on a patient. Meanwhile, deep neural networks have achieved superior performance in a wide variety of natural language processing (NLP) tasks because of their capacity to encode meaningful but abstract representations and learn the entire task end-to-end. In this thesis, I investigate representation learning of clinical narratives with deep neural networks through a number of tasks ranging from clinical concept extraction, clinical note modeling, and patient-level language representation. I present methods utilizing representation learning with neural networks to support understanding of clinical text documents. I first introduce the notion of representation learning from natural language processing and patient data modeling. Then, I investigate word-level representation learning to improve clinical concept extraction from clinical notes. I present two works on learning word representations and evaluate them to extract important concepts from clinical notes. The first study focuses on cancer-related information, and the second study evaluates shared-task data. The aims of these two studies are to automatically extract important entities from clinical notes. Next, I present a series of deep neural networks to encode hierarchical, longitudinal, and contextual information for modeling a series of clinical notes. I also evaluate the models by predicting clinical outcomes of interest, including mortality, length of stay, and phenotype predictions. Finally, I propose a novel representation learning architecture to develop a generalized and transferable language representation at the patient level. I also identify pre-training tasks appropriate for constructing a generalizable language representation. The main focus is to improve predictive performance of phenotypes with limited data, a challenging task due to a lack of data. Overall, this dissertation addresses issues in natural language processing for medicine, including clinical text classification and modeling. These studies show major barriers to understanding large-scale clinical notes. It is believed that developing deep representation learning methods for distilling enormous amounts of heterogeneous data into patient-level language representations will improve evidence-based clinical understanding. The approach to solving these issues by learning representations could be used across clinical applications despite noisy data. I conclude that considering different linguistic components in natural language and sequential information between clinical events is important. Such results have implications beyond the immediate context of predictions and further suggest future directions for clinical machine learning research to improve clinical outcomes. This could be a starting point for future phenotyping methods based on natural language processing that construct patient-level language representations to improve clinical predictions. While significant progress has been made, many open questions remain, so I will highlight a few works to demonstrate promising directions
    • …
    corecore