4 research outputs found

    Automatic BIRAD scoring of breast cancer mammograms

    Get PDF
    A computer aided diagnosis system (CAD) is developed to fully characterize and classify mass to benign and malignancy and to predict BIRAD (Breast Imaging Reporting and Data system) scores using mammographic image data. The CAD includes a preprocessing step to de-noise mammograms. This is followed by an active counter segmentation to deforms an initial curve, annotated by a radiologist, to separate and define the boundary of a mass from background. A feature extraction scheme wasthen used to fully characterize a mass by extraction of the most relevant features that have a large impact on the outcome of a patient biopsy. For this thirty-five medical and mathematical features based on intensity, shape and texture associated to the mass were extracted. Several feature selection schemes were then applied to select the most dominant features for use in next step, classification. Finally, a hierarchical classification schemes were applied on those subset of features to firstly classify mass to benign (mass with BIRAD score 2) and malignant mass (mass with BIRAD score over 4), and secondly to sub classify mass with BIRAD score over 4 to three classes (BIRAD with score 4a,4b,4c). Accuracy of segmentation performance were evaluated by calculating the degree of overlapping between the active counter segmentation and the manual segmentation, and the result was 98.5%. Also reproducibility of active counter 3 using different manual initialization of algorithm by three radiologists were assessed and result was 99.5%. Classification performance was evaluated using one hundred sixty masses (80 masses with BRAD score 2 and 80 mass with BIRAD score over4). The best result for classification of data to benign and malignance was found using a combination of sequential forward floating feature (SFFS) selection and a boosted tree hybrid classifier with Ada boost ensemble method, decision tree learner type and 100 learners’ regression tree classifier, achieving 100% sensitivity and specificity in hold out method, 99.4% in cross validation method and 98.62 % average accuracy in cross validation method. For further sub classification of eighty malignance data with BIRAD score of over 4 (30 mass with BIRAD score 4a,30 masses with BIRAD score 4b and 20 masses with BIRAD score 4c), the best result achieved using the boosted tree with ensemble method bag, decision tree learner type with 200 learners Classification, achieving 100% sensitivity and specificity in hold out method, 98.8% accuracy and 98.41% average accuracy for ten times run in cross validation method. Beside those 160 masses (BIRAD score 2 and over 4) 13 masses with BIRAD score 3 were gathered. Which means patient is recommended to be tested in another medical imaging technique and also is recommended to do follow-up in six months. The CAD system was trained with mass with BIRAD score 2 and over 4 also 4 it was further tested using 13 masses with a BIRAD score of 3 and the CAD results are shown to agree with the radiologist’s classification after confirming in six months follow up. The present results demonstrate high sensitivity and specificity of the proposed CAD system compared to prior research. The present research is therefore intended to make contributions to the field by proposing a novel CAD system, consists of series of well-selected image processing algorithms, to firstly classify mass to benign or malignancy, secondly sub classify BIRAD 4 to three groups and finally to interpret BIRAD 3 to BIRAD 2 without a need of follow up study

    Classification of mammographic lesions into BI-RADS shape categories using the beamlet transform

    No full text

    Pixel N-grams for Mammographic Image Classification

    Get PDF
    X-ray screening for breast cancer is an important public health initiative in the management of a leading cause of death for women. However, screening is expensive if mammograms are required to be manually assessed by radiologists. Moreover, manual screening is subject to perception and interpretation errors. Computer aided detection/diagnosis (CAD) systems can help radiologists as computer algorithms are good at performing image analysis consistently and repetitively. However, image features that enhance CAD classification accuracies are necessary for CAD systems to be deployed. Many CAD systems have been developed but the specificity and sensitivity is not high; in part because of challenges inherent in identifying effective features to be initially extracted from raw images. Existing feature extraction techniques can be grouped under three main approaches; statistical, spectral and structural. Statistical and spectral techniques provide global image features but often fail to distinguish between local pattern variations within an image. On the other hand, structural approach have given rise to the Bag-of-Visual-Words (BoVW) model, which captures local variations in an image, but typically do not consider spatial relationships between the visual “words”. Moreover, statistical features and features based on BoVW models are computationally very expensive. Similarly, structural feature computation methods other than BoVW are also computationally expensive and strongly dependent upon algorithms that can segment an image to localize a region of interest likely to contain the tumour. Thus, classification algorithms using structural features require high resource computers. In order for a radiologist to classify the lesions on low resource computers such as Ipads, Tablets, and Mobile phones, in a remote location, it is necessary to develop computationally inexpensive classification algorithms. Therefore, the overarching aim of this research is to discover a feature extraction/image representation model which can be used to classify mammographic lesions with high accuracy, sensitivity and specificity along with low computational cost. For this purpose a novel feature extraction technique called ‘Pixel N-grams’ is proposed. The Pixel N-grams approach is inspired from the character N-gram concept in text categorization. Here, N number of consecutive pixel intensities are considered in a particular direction. The image is then represented with the help of histogram of occurrences of the Pixel N-grams in an image. Shape and texture of mammographic lesions play an important role in determining the malignancy of the lesion. It was hypothesized that the Pixel N-grams would be able to distinguish between various textures and shapes. Experiments carried out on benchmark texture databases and binary basic shapes database have demonstrated that the hypothesis was correct. Moreover, the Pixel N-grams were able to distinguish between various shapes irrespective of size and location of shape in an image. The efficacy of the Pixel N-gram technique was tested on mammographic database of primary digital mammograms sourced from a radiological facility in Australia (LakeImaging Pty Ltd) and secondary digital mammograms (benchmark miniMIAS database). A senior radiologist from LakeImaging provided real time de-identified high resolution mammogram images with annotated regions of interests (which were used as groundtruth), and valuable radiological diagnostic knowledge. Two types of classifications were observed on these two datasets. Normal/abnormal classification useful for automated screening and circumscribed/speculation/normal classification useful for automated diagnosis of breast cancer. The classification results on both the mammography datasets using Pixel N-grams were promising. Classification performance (Fscore, sensitivity and specificity) using Pixel N-gram technique was observed to be significantly better than the existing techniques such as intensity histogram, co-occurrence matrix based features and comparable with the BoVW features. Further, Pixel N-gram features are found to be computationally less complex than the co-occurrence matrix based features as well as BoVW features paving the way for mammogram classification on low resource computers. Although, the Pixel N-gram technique was designed for mammographic classification, it could be applied to other image classification applications such as diabetic retinopathy, histopathological image classification, lung tumour detection using CT images, brain tumour detection using MRI images, wound image classification and tooth decay classification using dentistry x-ray images. Further, texture and shape classification is also useful for classification of real world images outside the medical domain. Therefore, the pixel N-gram technique could be extended for applications such as classification of satellite imagery and other object detection tasks.Doctor of Philosoph
    corecore