1,012 research outputs found

    Advanced Air Quality Management with Machine Learning

    Get PDF
    Air pollution has been a significant health risk factor at a regional and global scale. Although the present method can provide assessment indices like exposure risks or air pollutant concentrations for air quality management, the modeling estimations still remain non-negligible bias which could deviate from reality and limit the effectiveness of emission control strategies to reduce air pollution and derive health benefits. The current development in air quality management is still impeded by two major obstacles: (1) biased air quality concentrations from air quality models and (2) inaccurate exposure risk estimations Inspired by more available and overwhelming data, machine learning techniques provide promising opportunities to solve the above-mentioned obstacles and bridge the gap between model results and reality. This dissertation illustrates three machine learning applications to strengthen air quality management: (1) identifying heterogeneous exposure risk to air pollutants among diverse urbanization levels, (2) correcting modeled air pollutant concentrations and quantifying the bias of sources from model inputs, and (3) examine nonlinear air pollutant responses to local emissions. This dissertation uses Taiwan as a case study, due to its well-established hospital data, emission inventory, and air quality monitoring network. In conclusion, although ML models have become common in atmospheric and environmental health science in recent years, the modeling processes and output interpretation should rely on interdisciplinary professions and judgment. Except for meeting the basic modeling performance, future ML applications in atmospheric and environmental health science should provide interpretability and explainability in terms of human-environment interactions and interpretable physical/chemical mechanisms. Such applications are expected to feedback to traditional methods and deepen our understanding of environmental science

    Identification of significant factors for air pollution levels using a neural network based knowledge discovery system

    Get PDF
    Artificial neural network (ANN) is a commonly used approach to estimate or forecast air pollution levels, which are usually assessed by the concentrations of air contaminants such as nitrogen dioxide, sulfur dioxide, carbon monoxide, ozone, and suspended particulate matters (PMs) in the atmosphere of the concerned areas. Even through ANN can accurately estimate air pollution levels they are numerical enigmas and unable to provide explicit knowledge of air pollution levels by air pollution factors (e.g. traffic and meteorological factors). This paper proposed a neural network based knowledge discovery system aimed at overcoming this limitation in ANN. The system consists of two units: a) an ANN unit, which is used to estimate the air pollution levels based on relevant air pollution factors; b) a knowledge discovery unit, which is used to extract explicit knowledge from the ANN unit. To demonstrate the practicability of this neural network based knowledge discovery system, numerical data on mass concentrations of PM2.5 and PM1.0, meteorological and traffic data measured near a busy traffic road in Hangzhou city were applied to investigate the air pollution levels and the potential air pollution factors that may impact on the concentrations of these PMs. Results suggest that the proposed neural network based knowledge discovery system can accurately estimate air pollution levels and identify significant factors that have impact on air pollution levels

    HUMAN EXPOSURE TO AIR TOXICS IN URBAN ENVIRONMENTS: HEALTH RISKS, SOCIODEMOGRAPHIC DISPARITIES, AND MIXTURE PROFILES

    Get PDF
    Exposure to air toxics in urban environments may be of significant health concern because populations and emission sources are concentrated in the same geographic area. The overall objective of this study is to characterize the sources, variations, and mixture profiles of ambient air toxics in urban environments, and examine the sociodemographic disparities in exposures to air toxics in a typical U.S. metropolitan area.A model-to-monitor comparison was performed to evaluate the validity of modeling air toxics data using national datasets. Modeled concentrations in the 2011 National-scale Air Toxics Assessment (NATA) moderately agreed with monitoring measurements, and a sizable portion showed underestimation. Results warranted the need for actual monitoring data to conduct air toxics exposure assessment.Air toxics samples were collected in 106 census tracts in the Memphis area in 2014, and samples were analyzed for 71 volatile organic compounds (VOCs). Ambient VOC levels in Memphis were generally higher than the national averages in urban settings, but were mostly below the reference concentrations (RfCs). Factor analysis identified 5 major sources: manufacturing processes, vehicle exhaust, industrial solvents, refrigerants, and gasoline additives. The major non-cancer risks were from neurological, respiratory, and reproductive/developmental effects. The cumulative cancer risk was 5.9 3.3 10-4, with naphthalene and benzyl chloride as risk drivers.Sociodemographic disparities in cancer risks were examined by regressing cancer risks against socioeconomic, racial, and spatial parameters at the census tract level. We conducted separate disparity analyses using modeling data from 2011 NATA and our air toxics monitoring data. Analysis using modeling data showed strong sociodemographic disparities but that using monitoring data did not show. The discrepancy brought cautions for use of modeling air pollution data in environmental disparity research.We further assessed exposure to VOCs mixtures in five typical microenvironments (MEs): home, office, vehicle cabin, gas station, and community outdoors. The multivariate analysis of variance and pairwise analysis showed VOC profiles were distinguishingly different among MEs. The classification of profiles was achieved using the random forest. We anticipate wide applications of exposure profiles in epidemiologic research of exposure to air toxic mixtures

    Machine learning in geosciences and remote sensing

    Full text link
    Learning incorporates a broad range of complex procedures. Machine learning (ML) is a subdivision of artificial intelligence based on the biological learning process. The ML approach deals with the design of algorithms to learn from machine readable data. ML covers main domains such as data mining, difficult-to-program applications, and software applications. It is a collection of a variety of algorithms (e.g. neural networks, support vector machines, self-organizing map, decision trees, random forests, case-based reasoning, genetic programming, etc.) that can provide multivariate, nonlinear, nonparametric regression or classification. The modeling capabilities of the ML-based methods have resulted in their extensive applications in science and engineering. Herein, the role of ML as an effective approach for solving problems in geosciences and remote sensing will be highlighted. The unique features of some of the ML techniques will be outlined with a specific attention to genetic programming paradigm. Furthermore, nonparametric regression and classification illustrative examples are presented to demonstrate the efficiency of ML for tackling the geosciences and remote sensing problems

    Extended Quantitative Computed Tomography Analysis of Lung Structure and Function

    Get PDF
    Computed tomography (CT) imaging and quantitative CT (QCT) analysis for the study of lung health and disease have been rapidly advanced during the past decades, along with the employment of CT-based computational fluid dynamics (CFD) and machine learning approaches. The work presented in this thesis was devoted to extending the QCT analysis framework from three different perspectives.First, to extend the advanced QCT analysis to more data with undesirably protocolized CT scans, we developed a new deep learning-based automated segmentation of pulmonary lobes, in- corporating z-axis information into the conventional UNet segmentation. The proposed deep learn- ing segmentation, named ZUNet, was successfully applied for QCT analysis of silicosis patients with thick (5 or 10 mm) slices, which used to be excluded in QCT analysis since three-dimensional (3D) volumetric segmentation of the lungs and lobes were hardly successful or not automated. ZUNet outperformed UNet in lobe segmentation of human lungs. In addition, we extended the application of the QCT framework, combining CFD simulations for the entire subjects of the QCT analysis. One-dimensional (1D) CFD simulations of tidal breath- ing have been added to the inspiratory-expiratory CT image matching analysis of 66 asthma pa- tients (M:F=23:43, age=64.4±10.7) for pre- and post-bronchodilator comparison. We aimed to characterize comprehensive airway and lung structure and function relationship in the entire group response and patient-specific response to the bronchodilator. Along with the evidence of large air- way dilatation in the entire asthmatics, the CFD analysis revealed that improvements in regional flow rate fraction, particularly in the right lower lobe (RLL), airway pressure drop, airway resis- tance, and workload of breathing were significantly associated with the degree of large airway dilatation. Finally, we extended the approach using machine learning analysis to integrate numerous QCT variables with clinical features and additional information such as environmental exposure. In pursuit of investigating the effects of particulate matter (PM) exposure on human lung struc- ture and function alteration, principal component analysis (PCA) and k-means clustering iden- tified low, mid, and high exposure groups from directly measured air pollution exposure data of 270 healthy (age=68±10, M:F=15:51), asthma (age=60±12, M:F=39:56), chronic obstructive pulmonary disease (COPD) (age=69±7, M:F=66:10), and idiopathic pulmonary fibrosis (IPF) (age=72±7, M:F=43:10) subjects. Based on the exposure clusters, the RLL segmental airway narrowing was observed in the high exposure group. Various associations were found between the exposure data and about 200 multiscale lung features, from quantitative inspiratory and ex- piratory CT image matching and 1D CFD tidal breathing simulations. To highlight, small PM increases small airway disease in asthma. PM at all sizes decreases inspiratory low attenuation area in COPD and diseases luminal diameter of the RLL segmental airways in IPF

    Big data analytics for preventive medicine

    Get PDF
    © 2019, Springer-Verlag London Ltd., part of Springer Nature. Medical data is one of the most rewarding and yet most complicated data to analyze. How can healthcare providers use modern data analytics tools and technologies to analyze and create value from complex data? Data analytics, with its promise to efficiently discover valuable pattern by analyzing large amount of unstructured, heterogeneous, non-standard and incomplete healthcare data. It does not only forecast but also helps in decision making and is increasingly noticed as breakthrough in ongoing advancement with the goal is to improve the quality of patient care and reduces the healthcare cost. The aim of this study is to provide a comprehensive and structured overview of extensive research on the advancement of data analytics methods for disease prevention. This review first introduces disease prevention and its challenges followed by traditional prevention methodologies. We summarize state-of-the-art data analytics algorithms used for classification of disease, clustering (unusually high incidence of a particular disease), anomalies detection (detection of disease) and association as well as their respective advantages, drawbacks and guidelines for selection of specific model followed by discussion on recent development and successful application of disease prevention methods. The article concludes with open research challenges and recommendations

    Prediction Of Heart Failure Decompensations Using Artificial Intelligence - Machine Learning Techniques

    Get PDF
    Los apartados 4.41, 4.4.2 y 4.4.3 del capítulo 4 están sujetos a confidencialidad por la autora. 203 p.Heart failure (HF) is a major concern in public health. Its total impact is increased by its high incidence and prevalence and its unfavourable medium-term prognosis. In addition, HF leads to huge health care resource consumption. Moreover, efforts to develop a deterministic understanding of rehospitalization have been difficult, as no specific patient or hospital factors have been shown to consistently predict 30-day readmission after hospitalization for HF.Taking all these facts into account, we wanted to develop a project to improve the assistance care of patients with HF. Up to know, we were using telemonitoring with a codification system that generated alarms depending on the received values. However, these simple rules generated large number of false alerts being, hence, not trustworthy. The final aims of this work are: (i) asses the benefits of remote patient telemonitoring (RPT), (ii) improve the results obtained with RPT using ML techniques, detecting which parameters measured by telemonitoring best predict HF decompensations and creating predictive models that will reduce false alerts and detect early decompensations that otherwise will lead to hospital admissions and (iii) determine the influence of environmental factors on HF decompensations.All in all, the conclusions of this study are:1. Asses the benefits of RPT: Telemonitoring has not shown a statistically significant reduction in the number of HF-related hospital admissions. Nevertheless, we have observed a statistically significant reduction in mortality in the intervention group with a considerable percentage of deaths from non-cardiovascular causes. Moreover, patients have considered the RPT programme as a tool that can help them in the control of their chronic disease and in the relationship with health professionals.2. Improve the results obtained with RPT using machine learning techniques: Significant weight increases, desaturation below 90%, perception of clinical worsening, including development of oedema, worsening of functional class and orthopnoea are good predictors of heart failure decompensation. In addition, machine learning techniques have improved the current alerts system implemented in our hospital. The system reduces the number of false alerts notably although it entails a decrement on sensitivity values. The best results are achieved with the predictive model built by applying NB with Bernoulli to the combination of telemonitoring alerts and questionnaire alerts (Weight + Ankle + well-being plus the yellow alerts of systolic blood pressure, diastolic blood pressure, O2Sat and heart rate). 3. Determine the influence of environmental factors on HF decompensations: Air temperature is the most significant environmental factor (negative correlation) in our study, although some other attributes, such as precipitation, are also relevant. This work also shows a consistent association between increasing levels SO2 and NOX air and HF hospitalizations

    Contributions from computational intelligence to healthcare data processing

    Get PDF
    80 p.The increasing ability to gather, store and process health care information, through the electronic health records and improved communication methods opens the door for new applications intended to improve health care in many different ways. Crucial to this evolution is the development of new computational intelligence tools, related to machine learning and statistics. In this thesis we have dealt with two case studies involving health data. The first is the monitoring of children with respiratory diseases in the pediatric intensive care unit of a hospital. The alarm detection is stated as a classification problem predicting the triage selected by the nurse or medical doctor. The second is the prediction of readmissions leading to hospitalization in an emergency department of a hospital. Both problems have great impact in economic and personal well being. We have tackled them with a rigorous methodological approach, obtaining results that may lead to a real life implementation. We have taken special care in the treatment of the data imbalance. Finally we make propositions to bring these techniques to the clinical environment
    corecore