438 research outputs found

    Texture feature ranking with relevance learning to classify interstitial lung disease patterns

    Get PDF
    Objective: The generalized matrix learning vector quantization (GMLVQ) is used to estimate the relevance of texture features in their ability to classify interstitial lung disease patterns in high-resolution computed tomography images. Methodology: After a stochastic gradient descent, the GMLVQ algorithm provides a discriminative distance measure of relevance factors, which can account for pairwise correlations between different texture features and their importance for the classification of healthy and diseased patterns. 65 texture features were extracted from gray-level co-occurrence matrices (GLCMs). These features were ranked and selected according to their relevance obtained by GMLVQ and, for comparison, to a mutual information (MI) criteria. The classification performance for different feature subsets was calculated for a k-nearest-neighbor (kNN) and a random forests classifier (RanForest), and support vector machines with a linear and a radial basis function kernel (SVMlin and SVMrbf). Results: For all classifiers, feature sets selected by the relevance ranking assessed by GMLVQ had a significantly better classification performance (p <0.05) for many texture feature sets compared to the MI approach. For kNN, RanForest, and SVMrbf, some of these feature subsets had a significantly better classification performance when compared to the set consisting of all features (p <0.05). Conclusion: While this approach estimates the relevance of single features, future considerations of GMLVQ should include the pairwise correlation for the feature ranking, e.g. to reduce the redundancy of two equally relevant features. (C) 2012 Elsevier B.V. All rights reserved

    Optimization Strategies for Interactive Classification of Interstitial Lung Disease Textures

    Get PDF
    For computerized analysis of textures in interstitial lung disease, manual annotations of lung tissue are necessary. Since making these annotations is labor intensive, we previously proposed an interactive annotation framework. In this framework, observers iteratively trained a classifier to distinguish the different texture types by correcting its classification errors. In this work, we investigated three ways to extend this approach, in order to decrease the amount of user interaction required to annotate all lung tissue in a computed tomography scan. First, we conducted automatic classification experiments to test how data from previously annotated scans can be used for classification of the scan under consideration. We compared the performance of a classifier trained on data from one observer, a classifier trained on data from multiple observers, a classifier trained on consensus training data, and an ensemble of classifiers, each trained on data from different sources. Experiments were conducted without and with texture selection (ts). In the former case, training data from all eight textures was used. In the latter, only training data from the texture types present in the scan were used, and the observer would have to indicate textures contained in the scan to be analyzed. Second, we simulated interactive annotation to test the effects of (1) asking observers to perform ts before the start of annotation, (2) the use of a classifier trained on data from previously annotated scans at the start of annotation, when the interactive classifier is untrained, and (3) allowing observers to choose which interactive or automatic classification results they wanted to correct. Finally, various strategies for selecting the classification results that were presented to the observer were considered. Classification accuracies for all possible interactive annotation scenarios were compared. Using the best-performing protocol, in which observers select the textures that should be distinguished in the scan and in which they can choose which classification results to use for correction, a median accuracy of 88% was reached. The results obtained using this protocol were significantly better than results obtained with other interactive or automatic classification protocols

    Prediction of lung tumor types based on protein attributes by machine learning algorithms

    Full text link

    Differentiating Peripherally-Located Small Cell Lung Cancer From Non-small Cell Lung Cancer Using a CT Radiomic Approach

    Get PDF
    Lung cancer can be classified into two main categories: small cell lung cancer (SCLC) and non-small cell lung cancer (NSCLC), which are different in treatment strategy and survival probability. The lung CT images of SCLC and NSCLC are similar such that their subtle differences are hardly visually discernible by the human eye through conventional imaging evaluation. We hypothesize that SCLC/NSCLC differentiation could be achieved via computerized image feature analysis and classification in feature space, as termed a radiomic model. The purpose of this study was to use CT radiomics to differentiate SCLC from NSCLC adenocarcinoma. Patients with primary lung cancer, either SCLC or NSCLC adenocarcinoma, were retrospectively identified. The post-diagnosis pre-treatment lung CT images were used to segment the lung cancers. Radiomic features were extracted from histogram-based statistics, textural analysis of tumor images and their wavelet transforms. A minimal-redundancy-maximal-relevance method was used for feature selection. The predictive model was constructed with a multilayer artificial neural network. The performance of the SCLC/NSCLC adenocarcinoma classifier was evaluated by the area under the receiver operating characteristic curve (AUC). Our study cohort consisted of 69 primary lung cancer patients with SCLC (n = 35; age mean ± SD = 66.91± 9.75 years), and NSCLC adenocarcinoma (n = 34; age mean ± SD = 58.55 ± 11.94 years). The SCLC group had more male patients and smokers than the NSCLC group (P \u3c 0.05). Our SCLC/NSCLC classifier achieved an overall performance of AUC of 0.93 (95% confidence interval = [0.85, 0.97]), sensitivity = 0.85, and specificity = 0.85). Adding clinical data such as smoking history could improve the performance slightly. The top ranking radiomic features were mostly textural features. Our results showed that CT radiomics could quantitatively represent tumor heterogeneity and therefore could be used to differentiate primary lung cancer subtypes with satisfying results. CT image processing with the wavelet transformation technique enhanced the radiomic features for SCLC/NSCLC classification. Our pilot study should motivate further investigation of radiomics as a non-invasive approach for early diagnosis and treatment of lung cancer

    Predictive based hybrid ranker to yield significant features in writer identification

    Get PDF
    The contribution of writer identification (WI) towards personal identification in biometrics traits is known because it is easily accessible, cheaper, more reliable and acceptable as compared to other methods such as personal identification based DNA, iris and fingerprint. However, the production of high dimensional datasets has resulted into too many irrelevant or redundant features. These unnecessary features increase the size of the search space and decrease the identification performance. The main problem is to identify the most significant features and select the best subset of features that can precisely predict the authors. Therefore, this study proposed the hybridization of GRA Features Ranking and Feature Subset Selection (GRAFeSS) to develop the best subsets of highest ranking features and developed discretization model with the hybrid method (Dis-GRAFeSS) to improve classification accuracy. Experimental results showed that the methods improved the performance accuracy in identifying the authorship of features based ranking invariant discretization by substantially reducing redundant features

    Texture Analysis and Machine Learning to Predict Pulmonary Ventilation from Thoracic Computed Tomography

    Get PDF
    Chronic obstructive pulmonary disease (COPD) leads to persistent airflow limitation, causing a large burden to patients and the health care system. Thoracic CT provides an opportunity to observe the structural pathophysiology of COPD, whereas hyperpolarized gas MRI provides images of the consequential ventilation heterogeneity. However, hyperpolarized gas MRI is currently limited to research centres, due to the high cost of gas and polarization equipment. Therefore, I developed a pipeline using texture analysis and machine learning methods to create predicted ventilation maps based on non-contrast enhanced, single-volume thoracic CT. In a COPD cohort, predicted ventilation maps were qualitatively and quantitatively related to ground-truth MRI ventilation, and both maps were related to important patient lung function and quality-of-life measures. This study is the first to demonstrate the feasibility of predicting hyperpolarized MRI-based ventilation from single-volume, breath-hold thoracic CT, which has potential to translate pulmonary ventilation information to widely available thoracic CT imaging
    corecore