15 research outputs found

    Signal processing for automatic heartbeat classification and patient adaptation in the electrocardiogram

    Get PDF
    Las enfermedades cardiovasculares son en la actualidad la mayor causa de muerte individual en los países desarrollados, por lo tanto cualquier avance en las metodologías para el diagnóstico podrían mejorar la salud de muchas personas. Dentro de las enfermedades cardiovasculares, la muerte súbita cardíaca es una de las causas de muerte más importantes, por su número y por el impacto social que provoca. Sin lugar a duda se trata uno de los grandes desafíos de la cardiología moderna. Hay evidencias para relacionar las arritmias con la muerte súbita cardíaca. Por otro lado, la clasificación de latidos en el electrocardiograma (ECG) es un análisis previo para el estudio de las arritmias. El análisis del ECG proporciona una técnica no invasiva para el estudio de la actividad del corazón en sus distintas condiciones. Particularmente los algoritmos automáticos de clasificación se focalizan en el análisis del ritmo y la morfología del ECG, y específicamente en las variaciones respecto a la normalidad. Justamente, las variaciones en el ritmo, regularidad, lugar de origen y forma de conducción de los impulsos cardíacos, se denominan arritmias. Mientras que algunas arritmias representan una amenaza inminente (Ej. fibrilación ventricular), existen otras más sutiles que pueden ser una amenaza a largo plazo sin el tratamiento adecuado. Es en estos últimos casos, que registros ECG de larga duración requieren una inspección cuidadosa, donde los algoritmos automáticos de clasificación representan una ayuda significativa en el diagnóstico. En la última década se han desarrollado algunos algoritmos de clasificación de ECG, pero solo unos pocos tienen metodologías y resultados comparables, a pesar de las recomendaciones de la AAMI para facilitar la resolución de estos problemas. De dichos métodos, algunos funcionan de manera completamente automática, mientras que otros pueden aprovechar la asistencia de un experto para mejorar su desempeño. La base de datos utilizada en todos estos trabajos ha sido la MIT-BIH de arritmias. En cuanto a las características utilizadas, los intervalos RR fueron usados por casi todos los grupos. También se utilizaron muestras del complejo QRS diezmado, o transformado mediante polinomios de Hermite, transformada de Fourier o la descomposición wavelet. Otros grupos usaron características que integran la información presente en ambas derivaciones, como el máximo del vectocardiograma del complejo QRS, o el ángulo formado en dicho punto. El objetivo de esta tesis ha sido estudiar algunas metodologías para la clasificación de latidos en el ECG. En primer lugar se estudiaron metodologías automáticas, con capacidad para contemplar el análisis de un número arbitrario de derivaciones. Luego se estudió la adaptación al paciente y la posibilidad de incorporar la asistencia de un experto para mejorar el rendimiento del clasificador automático. En principio se desarrolló y validó un clasificador de latidos sencillo, que utiliza características seleccionadas en base a una buena capacidad de generalización. Se han considerado características de la serie de intervalos RR (distancia entre dos latidos consecutivos), como también otras calculadas a partir de ambas derivaciones de la señal de ECG, y escalas de su transformada wavelet. Tanto el desempeño en la clasificación como la capacidad de generalización han sido evaluados en bases de datos públicas: la MIT-BIH de arritmias, la MIT-BIH de arritmias supraventriculares y la del Instituto de Técnicas Cardiológicas de San Petersburgo (INCART). Se han seguido las recomendaciones de la Asociación para el Avance de la Instrumentación Médica (AAMI) tanto para el etiquetado de clases como para la presentación de los resultados. Para la búsqueda de características se adoptó un algoritmo de búsqueda secuencial flotante, utilizando diferentes criterios de búsqueda, para luego elegir el modelo con mejor rendimiento y capacidad de generalización en los sets de entrenamiento y validación. El mejor modelo encontrado incluye 8 características y ha sido entrenado y evaluado en particiones disjuntas de la MIT-BIH de arritmias. Todas las carácterísticas del modelo corresponden a mediciones de intervalos temporales. Esto puede explicarse debido a que los registros utilizados en los experimentos no siempre contienen las mismas derivaciones, y por lo tanto la capacidad de clasificación de aquellas características basadas en amplitudes se ve seriamente disminuida. Las primeras 4 características del modelo están claramente relacionadas a la evolución del ritmo cardíaco, mientras que las otras cuatro pueden interpretarse como mediciones alternativas de la anchura del complejo QRS, y por lo tanto morfológicas. Como resultado, el modelo obtenido tiene la ventaja evidente de un menor tamaño, lo que redunda tanto en un ahorro computacional como en una mejor estimación de los parámetros del modelo durante el entrenamiento. Como ventaja adicional, este modelo depende exclusivamente de la detección de cada latido, haciendo este clasificador especialmente útil en aquellos casos donde la delineación de las ondas del ECG no puede realizarse de manera confiable. Los resultados obtenidos en el set de evaluación han sido: exactitud global (A) de 93%; para latidos normales: sensibilidad (S) 95% valor predictivo positivo (P^{+}) 98%; para latidos supraventriculares, S 77%, P^{+} 39%; y para latidos ventriculares S 81%, P^{+} 87%. Para comprobar la capacidad de generalización, se evaluó el rendimiento en la INCART obteniéndose resultados comparables a los del set de evaluación. El modelo de clasificación obtenido utiliza menos características, y adicionalmente presentó mejor rendimiento y capacidad de generalización que otros representativos del estado del arte. Luego se han estudiado dos mejoras para el clasificador desarrollado en el párrafo anterior. La primera fue adaptarlo a registros ECG de un número arbitrario de derivaciones, o extensión multiderivacional. En la segunda mejora se buscó cambiar el clasificador lineal por un perceptrón multicapa no lineal (MLP). Para la extensión multiderivacional se estudió si conlleva alguna mejora incluir información del ECG multiderivacional en el modelo previamente validado. Dicho modelo incluye características calculadas de la serie de intervalos RR y descriptores morfológicos calculados en la transformada wavelet de cada derivación. Los experimentos se han realizado en la INCART, disponible en Physionet, mientras que la generalización se corroboró en otras bases de datos públicas y privadas. En todas las bases de datos se siguieron las recomendaciones de la AAMI para el etiquetado de clases y presentación de resultados. Se estudiaron varias estrategias para incorporar la información adicional presente en registros de 12 derivaciones. La mejor estrategia consistió en realizar el análisis de componentes principales a la transformada wavelet del ECG. El rendimiento obtenido con dicha estrategia fue: para latidos normales: S98%, P^{+}93%; para latidos supraventriculares, S86%, P^{+}91%; y para latidos ventriculares S90%, P^{+}90%. La capacidad de generalización de esta estrategia se comprobó tras evaluarla en otras bases de datos, con diferentes cantidades de derivaciones, obteniendo resultados comparables. En conclusión, se mejoró el rendimiento del clasificador de referencia tras incluir la información disponible en todas las derivaciones disponibles. La mejora del clasificador lineal por medio de un MLP se realizó siguiendo una metodología similar a la descrita más arriba. El rendimiento obtenido fue: A 89%; para latidos normales: S90%, P^{+}99% para latidos supraventriculares, S83%, P^{+}34%; para latidos ventriculares S87%, P^{+}76%. Finalmente estudiamos un algoritmo de clasificación basado en las metodologías descritas en los anteriores párrafos, pero con la capacidad de mejorar su rendimiento mediante la ayuda de un experto. Se presentó un algoritmo de clasificación de latidos en el ECG adaptable al paciente, basado en el clasificador automático previamente desarrollado y un algoritmo de clustering. Tanto el clasificador automático, como el algoritmo de clustering utilizan características calculadas de la serie de intervalos RR y descriptores de morfología calculados de la transformada wavelet. Integrando las decisiones de ambos clasificadores, este algoritmo puede desempeñarse automáticamente o con varios grados de asistencia. El algoritmo ha sido minuciosamente evaluado en varias bases de datos para facilitar la comparación. Aún en el modo completamente automático, el algoritmo mejora el rendimiento del clasificador automático original; y con menos de 2 latidos anotados manualmente (MAHB) por registro, el algoritmo obtuvo una mejora media para todas las bases de datos del 6.9% en A, de 6.5\%S y de 8.9\% en P^{+}. Con una asistencia de solo 12 MAHB por registro resultó en una mejora media de 13.1\%en A, de 13.9\% en S y de 36.1\% en P^{+}. En el modo asistido, el algoritmo obtuvo un rendimiento superior a otros representativos del estado del arte, con menor asistencia por parte del experto. Como conclusiones de la tesis, debemos enfatizar la etapa del diseño y análisis minucioso de las características a utilizar. Esta etapa está íntimamente ligada al conocimiento del problema a resolver. Por otro lado, la selección de un subset de características ha resultado muy ventajosa desde el punto de la eficiencia computacional y la capacidad de generalización del modelo obtenido. En último lugar, la utilización de un clasificador simple o de baja capacidad (por ejemplo funciones discriminantes lineales) asegurará que el modelo de características sea responsable en mayor parte del rendimiento global del sistema. Con respecto a los sets de datos para la realización de los experimentos, es fundamental contar con un elevado numero de sujetos. Es importante incidir en la importancia de contar con muchos sujetos, y no muchos registros de pocos sujetos, dada la gran variabilidad intersujeto observada. De esto se desprende la necesidad de evaluar la capacidad de generalización del sistema a sujetos no contemplados durante el entrenamiento o desarrollo. Por último resaltaremos la complejidad de comparar el rendimiento de clasificadores en problemas mal balanceados, es decir que las clases no se encuentras igualmente representadas. De las alternativas sugeridas en esta tesis probablemente la más recomendable sea la matriz de confusión, ya que brinda una visión completa del rendimiento del clasificador, a expensas de una alta redundancia. Finalmente, luego de realizar comparaciones justas con otros trabajos representativos del estado actual de la técnica, concluimos que los resultados presentados en esta tesis representan una mejora en el campo de la clasificación de latidos automática y adaptada al paciente, en la señal de ECG

    Extraction and Detection of Fetal Electrocardiograms from Abdominal Recordings

    Get PDF
    The non-invasive fetal ECG (NIFECG), derived from abdominal surface electrodes, offers novel diagnostic possibilities for prenatal medicine. Despite its straightforward applicability, NIFECG signals are usually corrupted by many interfering sources. Most significantly, by the maternal ECG (MECG), whose amplitude usually exceeds that of the fetal ECG (FECG) by multiple times. The presence of additional noise sources (e.g. muscular/uterine noise, electrode motion, etc.) further affects the signal-to-noise ratio (SNR) of the FECG. These interfering sources, which typically show a strong non-stationary behavior, render the FECG extraction and fetal QRS (FQRS) detection demanding signal processing tasks. In this thesis, several of the challenges regarding NIFECG signal analysis were addressed. In order to improve NIFECG extraction, the dynamic model of a Kalman filter approach was extended, thus, providing a more adequate representation of the mixture of FECG, MECG, and noise. In addition, aiming at the FECG signal quality assessment, novel metrics were proposed and evaluated. Further, these quality metrics were applied in improving FQRS detection and fetal heart rate estimation based on an innovative evolutionary algorithm and Kalman filtering signal fusion, respectively. The elaborated methods were characterized in depth using both simulated and clinical data, produced throughout this thesis. To stress-test extraction algorithms under ideal circumstances, a comprehensive benchmark protocol was created and contributed to an extensively improved NIFECG simulation toolbox. The developed toolbox and a large simulated dataset were released under an open-source license, allowing researchers to compare results in a reproducible manner. Furthermore, to validate the developed approaches under more realistic and challenging situations, a clinical trial was performed in collaboration with the University Hospital of Leipzig. Aside from serving as a test set for the developed algorithms, the clinical trial enabled an exploratory research. This enables a better understanding about the pathophysiological variables and measurement setup configurations that lead to changes in the abdominal signal's SNR. With such broad scope, this dissertation addresses many of the current aspects of NIFECG analysis and provides future suggestions to establish NIFECG in clinical settings.:Abstract Acknowledgment Contents List of Figures List of Tables List of Abbreviations List of Symbols (1)Introduction 1.1)Background and Motivation 1.2)Aim of this Work 1.3)Dissertation Outline 1.4)Collaborators and Conflicts of Interest (2)Clinical Background 2.1)Physiology 2.1.1)Changes in the maternal circulatory system 2.1.2)Intrauterine structures and feto-maternal connection 2.1.3)Fetal growth and presentation 2.1.4)Fetal circulatory system 2.1.5)Fetal autonomic nervous system 2.1.6)Fetal heart activity and underlying factors 2.2)Pathology 2.2.1)Premature rupture of membrane 2.2.2)Intrauterine growth restriction 2.2.3)Fetal anemia 2.3)Interpretation of Fetal Heart Activity 2.3.1)Summary of clinical studies on FHR/FHRV 2.3.2)Summary of studies on heart conduction 2.4)Chapter Summary (3)Technical State of the Art 3.1)Prenatal Diagnostic and Measuring Technique 3.1.1)Fetal heart monitoring 3.1.2)Related metrics 3.2)Non-Invasive Fetal ECG Acquisition 3.2.1)Overview 3.2.2)Commercial equipment 3.2.3)Electrode configurations 3.2.4)Available NIFECG databases 3.2.5)Validity and usability of the non-invasive fetal ECG 3.3)Non-Invasive Fetal ECG Extraction Methods 3.3.1)Overview on the non-invasive fetal ECG extraction methods 3.3.2)Kalman filtering basics 3.3.3)Nonlinear Kalman filtering 3.3.4)Extended Kalman filter for FECG estimation 3.4)Fetal QRS Detection 3.4.1)Merging multichannel fetal QRS detections 3.4.2)Detection performance 3.5)Fetal Heart Rate Estimation 3.5.1)Preprocessing the fetal heart rate 3.5.2)Fetal heart rate statistics 3.6)Fetal ECG Morphological Analysis 3.7)Problem Description 3.8)Chapter Summary (4)Novel Approaches for Fetal ECG Analysis 4.1)Preliminary Considerations 4.2)Fetal ECG Extraction by means of Kalman Filtering 4.2.1)Optimized Gaussian approximation 4.2.2)Time-varying covariance matrices 4.2.3)Extended Kalman filter with unknown inputs 4.2.4)Filter calibration 4.3)Accurate Fetal QRS and Heart Rate Detection 4.3.1)Multichannel evolutionary QRS correction 4.3.2)Multichannel fetal heart rate estimation using Kalman filters 4.4)Chapter Summary (5)Data Material 5.1)Simulated Data 5.1.1)The FECG Synthetic Generator (FECGSYN) 5.1.2)The FECG Synthetic Database (FECGSYNDB) 5.2)Clinical Data 5.2.1)Clinical NIFECG recording 5.2.2)Scope and limitations of this study 5.2.3)Data annotation: signal quality and fetal amplitude 5.2.4)Data annotation: fetal QRS annotation 5.3)Chapter Summary (6)Results for Data Analysis 6.1)Simulated Data 6.1.1)Fetal QRS detection 6.1.2)Morphological analysis 6.2)Own Clinical Data 6.2.1)FQRS correction using the evolutionary algorithm 6.2.2)FHR correction by means of Kalman filtering (7)Discussion and Prospective 7.1)Data Availability 7.1.1)New measurement protocol 7.2)Signal Quality 7.3)Extraction Methods 7.4)FQRS and FHR Correction Algorithms (8)Conclusion References (A)Appendix A - Signal Quality Annotation (B)Appendix B - Fetal QRS Annotation (C)Appendix C - Data Recording GU

    Computational Intelligence in Healthcare

    Get PDF
    This book is a printed edition of the Special Issue Computational Intelligence in Healthcare that was published in Electronic

    Computational Intelligence in Healthcare

    Get PDF
    The number of patient health data has been estimated to have reached 2314 exabytes by 2020. Traditional data analysis techniques are unsuitable to extract useful information from such a vast quantity of data. Thus, intelligent data analysis methods combining human expertise and computational models for accurate and in-depth data analysis are necessary. The technological revolution and medical advances made by combining vast quantities of available data, cloud computing services, and AI-based solutions can provide expert insight and analysis on a mass scale and at a relatively low cost. Computational intelligence (CI) methods, such as fuzzy models, artificial neural networks, evolutionary algorithms, and probabilistic methods, have recently emerged as promising tools for the development and application of intelligent systems in healthcare practice. CI-based systems can learn from data and evolve according to changes in the environments by taking into account the uncertainty characterizing health data, including omics data, clinical data, sensor, and imaging data. The use of CI in healthcare can improve the processing of such data to develop intelligent solutions for prevention, diagnosis, treatment, and follow-up, as well as for the analysis of administrative processes. The present Special Issue on computational intelligence for healthcare is intended to show the potential and the practical impacts of CI techniques in challenging healthcare applications

    Assessing Variability of EEG and ECG/HRV Time Series Signals Using a Variety of Non-Linear Methods

    Get PDF
    Time series signals, such as Electroencephalogram (EEG) and Electrocardiogram (ECG) represent the complex dynamic behaviours of biological systems. The analysis of these signals using variety of nonlinear methods is essential for understanding variability within EEG and ECG, which potentially could help unveiling hidden patterns related to underlying physiological mechanisms. EEG is a time varying signal, and electrodes for recording EEG at different positions on the scalp give different time varying signals. There might be correlation between these signals. It is important to know the correlation between EEG signals because it might tell whether or not brain activities from different areas are related. EEG and ECG might be related to each other because both of them are generated from one co-ordinately working body. Investigating this relationship is of interest because it may reveal information about the correlation between EEG and ECG signals. This thesis is about assessing variability of time series data, EEG and ECG, using variety of nonlinear measures. Although other research has looked into the correlation between EEGs using a limited number of electrodes and a limited number of combinations of electrode pairs, no research has investigated the correlation between EEG signals and distance between electrodes. Furthermore, no one has compared the correlation performance for participants with and without medical conditions. In my research, I have filled up these gaps by using a full range of electrodes and all possible combinations of electrode pairs analysed in Time Domain (TD). Cross-Correlation method is calculated on the processed EEG signals for different number unique electrode pairs from each datasets. In order to obtain the distance in centimetres (cm) between electrodes, a measuring tape was used. For most of our participants the head circumference range was 54-58cm, for which a medium-sized I have discovered that the correlation between EEG signals measured through electrodes is linearly dependent on the physical distance (straight-line) distance between them for datasets without medical condition, but not for datasets with medical conditions. Some research has investigated correlation between EEG and Heart Rate Variability (HRV) within limited brain areas and demonstrated the existence of correlation between EEG and HRV. But no research has indicated whether or not the correlation changes with brain area. Although Wavelet Transformations (WT) have been performed on time series data including EEG and HRV signals to extract certain features respectively by other research, so far correlation between WT signals of EEG and HRV has not been analysed. My research covers these gaps by conducting a thorough investigation of all electrodes on the human scalp in Frequency Domain (FD) as well as TD. For the reason of different sample rates of EEG and HRV, two different approaches (named as Method 1 and Method 2) are utilised to segment EEG signals and to calculate Pearson’s Correlation Coefficient for each of the EEG frequencies with each of the HRV frequencies in FD. I have demonstrated that EEG at the front area of the brain has a stronger correlation with HRV than that at the other area in a frequency domain. These findings are independent of both participants and brain hemispheres. Sample Entropy (SE) is used to predict complexity of time series data. Recent research has proposed new calculation methods for SE, aiming to improve the accuracy. To my knowledge, no one has attempted to reduce the computational time of SE calculation. I have developed a new calculation method for time series complexity which could improve computational time significantly in the context of calculating a correlation between EEG and HRV. The results have a parsimonious outcome of SE calculation by exploiting a new method of SE implementation. In addition, it is found that the electrical activity in the frontal lobe of the brain appears to be correlated with the HRV in a time domain. Time series analysis method has been utilised to study complex systems that appear ubiquitous in nature, but limited to certain dynamic systems (e.g. analysing variables affecting stock values). In this thesis, I have also investigated the nature of the dynamic system of HRV. I have disclosed that Embedding Dimension could unveil two variables that determined HRV

    Smart Sensors for Healthcare and Medical Applications

    Get PDF
    This book focuses on new sensing technologies, measurement techniques, and their applications in medicine and healthcare. Specifically, the book briefly describes the potential of smart sensors in the aforementioned applications, collecting 24 articles selected and published in the Special Issue “Smart Sensors for Healthcare and Medical Applications”. We proposed this topic, being aware of the pivotal role that smart sensors can play in the improvement of healthcare services in both acute and chronic conditions as well as in prevention for a healthy life and active aging. The articles selected in this book cover a variety of topics related to the design, validation, and application of smart sensors to healthcare

    Complexity Science in Human Change

    Get PDF
    This reprint encompasses fourteen contributions that offer avenues towards a better understanding of complex systems in human behavior. The phenomena studied here are generally pattern formation processes that originate in social interaction and psychotherapy. Several accounts are also given of the coordination in body movements and in physiological, neuronal and linguistic processes. A common denominator of such pattern formation is that complexity and entropy of the respective systems become reduced spontaneously, which is the hallmark of self-organization. The various methodological approaches of how to model such processes are presented in some detail. Results from the various methods are systematically compared and discussed. Among these approaches are algorithms for the quantification of synchrony by cross-correlational statistics, surrogate control procedures, recurrence mapping and network models.This volume offers an informative and sophisticated resource for scholars of human change, and as well for students at advanced levels, from graduate to post-doctoral. The reprint is multidisciplinary in nature, binding together the fields of medicine, psychology, physics, and neuroscience

    Echocardiography

    Get PDF
    The book "Echocardiography - New Techniques" brings worldwide contributions from highly acclaimed clinical and imaging science investigators, and representatives from academic medical centers. Each chapter is designed and written to be accessible to those with a basic knowledge of echocardiography. Additionally, the chapters are meant to be stimulating and educational to the experts and investigators in the field of echocardiography. This book is aimed primarily at cardiology fellows on their basic echocardiography rotation, fellows in general internal medicine, radiology and emergency medicine, and experts in the arena of echocardiography. Over the last few decades, the rate of technological advancements has developed dramatically, resulting in new techniques and improved echocardiographic imaging. The authors of this book focused on presenting the most advanced techniques useful in today's research and in daily clinical practice. These advanced techniques are utilized in the detection of different cardiac pathologies in patients, in contributing to their clinical decision, as well as follow-up and outcome predictions. In addition to the advanced techniques covered, this book expounds upon several special pathologies with respect to the functions of echocardiography

    Lung cancer: sex difference in the lifetime risk and 10-year risk between 1995 and 2013 in a Swiss population

    Get PDF
    Introduction: In Switzerland, lung cancer is a leading cause of cancer death. Because smoking is the major cause of lung cancer, trends in lung cancer incidence are following trends in smoking habits in the population, with a latency time of about 30 years. In Switzerland, there was a peak in men’s lung cancer incidence in the 1980s, followed by a decrease until now. Among women, the incidence has increased since the 1970s and, apparently, has not yet reached a peak. Because cancers are feared diseases, an adequate communication about the individual risk of developing cancer is important. Mortality and incidence are traditionally used to assess cancer burden. However, these metrics are difficult to interpret at the individual level. Providing the lifetime and 10-year risk of cancer could improve risk communication for patients and health professionals. Our aim was to estimate trends in the lifetime and 10-year risk of lung cancer, in men and women, between 1995 and 2013

    Is overdiagnosis of prostate cancer leveling off? Recent changes in incidence and surgery rates in Switzerland

    Get PDF
    Many western countries, including Switzerland. Various organizations have recently recommended against routine screening, notably due to the high risk of overdiagnosis or overtreatment. Our aim was to examine whether recent changes in secular trends in the incidence and mortality of prostate cancer, as well as prostatectomy rates have been observed in Switzerland
    corecore