29,632 research outputs found

    Extracting information from the text of electronic medical records to improve case detection: a systematic review

    Get PDF
    Background: Electronic medical records (EMRs) are revolutionizing health-related research. One key issue for study quality is the accurate identification of patients with the condition of interest. Information in EMRs can be entered as structured codes or unstructured free text. The majority of research studies have used only coded parts of EMRs for case-detection, which may bias findings, miss cases, and reduce study quality. This review examines whether incorporating information from text into case-detection algorithms can improve research quality. Methods: A systematic search returned 9659 papers, 67 of which reported on the extraction of information from free text of EMRs with the stated purpose of detecting cases of a named clinical condition. Methods for extracting information from text and the technical accuracy of case-detection algorithms were reviewed. Results: Studies mainly used US hospital-based EMRs, and extracted information from text for 41 conditions using keyword searches, rule-based algorithms, and machine learning methods. There was no clear difference in case-detection algorithm accuracy between rule-based and machine learning methods of extraction. Inclusion of information from text resulted in a significant improvement in algorithm sensitivity and area under the receiver operating characteristic in comparison to codes alone (median sensitivity 78% (codes + text) vs 62% (codes), P = .03; median area under the receiver operating characteristic 95% (codes + text) vs 88% (codes), P = .025). Conclusions: Text in EMRs is accessible, especially with open source information extraction algorithms, and significantly improves case detection when combined with codes. More harmonization of reporting within EMR studies is needed, particularly standardized reporting of algorithm accuracy metrics like positive predictive value (precision) and sensitivity (recall)

    Investigating SAR algorithm for spaceborne interferometric oil spill detection

    Get PDF
    The environmental damages and recovery of terrestrial ecosystems from oil spills can last decades. Oil spills have been responsible for loss of aquamarine lives, organisms, trees, vegetation, birds and wildlife. Although there are several methods through which oil spills can be detected, it can be argued that remote sensing via the use of spaceborne platforms provides enormous benefits. This paper will provide more efficient means and methods that can assist in improving oil spill responses. The objective of this research is to develop a signal processing algorithm that can be used for detecting oil spills using spaceborne SAR interferometry (InSAR) data. To this end, a pendulum formation of multistatic smallSAR carrying platforms in a near equatorial orbit is described. The characteristic parameters such as the effects of incidence angles on radar backscatter, which support the detection of oil spills, will be the main drivers for determining the relative positions of the small satellites in formation. The orbit design and baseline distances between each spaceborne SAR platform will also be discussed. Furthermore, results from previous analysis on coverage assessment and revisit time shall be highlighted. Finally, an evaluation of automatic algorithm techniques for oil spill detection in SAR images will be conducted and results presented. The framework for the automatic algorithm considered consists of three major steps. The segmentation stage, where techniques that suggest the use of thresholding for dark spot segmentation within the captured InSAR image scene is conducted. The feature extraction stage involves the geometry and shape of the segmented region where elongation of the oil slick is considered an important feature and a function of the width and the length of the oil slick. For the classification stage, where the major objective is to distinguish oil spills from look-alikes, a Mahalanobis classifier will be used to estimate the probability of the extracted features being oil spills. The validation process of the algorithm will be conducted by using NASA’s UAVSAR data obtained over the Gulf of coast oil spill and RADARSAT-1 dat

    Evaluating tools to support a new practical classification of diabetes: excellent control may represent misdiagnosis and omission from disease registers is associated with worse control.

    Get PDF
    To conduct a service evaluation of usability and utility on-line clinical audit tools developed as part of a UK Classification of Diabetes project to improve the categorisation and ultimately management of diabetes

    Application of spectral and spatial indices for specific class identification in Airborne Prism EXperiment (APEX) imaging spectrometer data for improved land cover classification

    Get PDF
    Hyperspectral remote sensing's ability to capture spectral information of targets in very narrow bandwidths gives rise to many intrinsic applications. However, the major limiting disadvantage to its applicability is its dimensionality, known as the Hughes Phenomenon. Traditional classification and image processing approaches fail to process data along many contiguous bands due to inadequate training samples. Another challenge of successful classification is to deal with the real world scenario of mixed pixels i.e. presence of more than one class within a single pixel. An attempt has been made to deal with the problems of dimensionality and mixed pixels, with an objective to improve the accuracy of class identification. In this paper, we discuss the application of indices to cope with the disadvantage of the dimensionality of the Airborne Prism EXperiment (APEX) hyperspectral Open Science Dataset (OSD) and to improve the classification accuracy using the Possibilistic c–Means (PCM) algorithm. This was used for the formulation of spectral and spatial indices to describe the information in the dataset in a lesser dimensionality. This reduced dimensionality is used for classification, attempting to improve the accuracy of determination of specific classes. Spectral indices are compiled from the spectral signatures of the target and spatial indices have been defined using texture analysis over defined neighbourhoods. The classification of 20 classes of varying spatial distributions was considered in order to evaluate the applicability of spectral and spatial indices in the extraction of specific class information. The classification of the dataset was performed in two stages; spectral and a combination of spectral and spatial indices individually as input for the PCM classifier. In addition to the reduction of entropy, while considering a spectral-spatial indices approach, an overall classification accuracy of 80.50% was achieved, against 65% (spectral indices only) and 59.50% (optimally determined principal component
    • …
    corecore