4 research outputs found
Automatic BIRAD scoring of breast cancer mammograms
A computer aided diagnosis system (CAD) is developed to fully characterize and
classify mass to benign and malignancy and to predict BIRAD (Breast Imaging
Reporting and Data system) scores using mammographic image data. The CAD
includes a preprocessing step to de-noise mammograms. This is followed by an
active counter segmentation to deforms an initial curve, annotated by a
radiologist, to separate and define the boundary of a mass from background. A
feature extraction scheme wasthen used to fully characterize a mass by extraction
of the most relevant features that have a large impact on the outcome of a patient
biopsy. For this thirty-five medical and mathematical features based on intensity,
shape and texture associated to the mass were extracted. Several feature selection
schemes were then applied to select the most dominant features for use in next
step, classification. Finally, a hierarchical classification schemes were applied on
those subset of features to firstly classify mass to benign (mass with BIRAD score
2) and malignant mass (mass with BIRAD score over 4), and secondly to sub classify
mass with BIRAD score over 4 to three classes (BIRAD with score 4a,4b,4c).
Accuracy of segmentation performance were evaluated by calculating the degree
of overlapping between the active counter segmentation and the manual
segmentation, and the result was 98.5%. Also reproducibility of active counter
3
using different manual initialization of algorithm by three radiologists were
assessed and result was 99.5%.
Classification performance was evaluated using one hundred sixty masses (80
masses with BRAD score 2 and 80 mass with BIRAD score over4). The best result
for classification of data to benign and malignance was found using a combination
of sequential forward floating feature (SFFS) selection and a boosted tree hybrid
classifier with Ada boost ensemble method, decision tree learner type and 100
learnersâ regression tree classifier, achieving 100% sensitivity and specificity in
hold out method, 99.4% in cross validation method and 98.62 % average accuracy
in cross validation method.
For further sub classification of eighty malignance data with BIRAD score of over
4 (30 mass with BIRAD score 4a,30 masses with BIRAD score 4b and 20 masses with
BIRAD score 4c), the best result achieved using the boosted tree with ensemble
method bag, decision tree learner type with 200 learners Classification, achieving
100% sensitivity and specificity in hold out method, 98.8% accuracy and 98.41%
average accuracy for ten times run in cross validation method.
Beside those 160 masses (BIRAD score 2 and over 4) 13 masses with BIRAD score
3 were gathered. Which means patient is recommended to be tested in another
medical imaging technique and also is recommended to do follow-up in six
months. The CAD system was trained with mass with BIRAD score 2 and over 4 also
4
it was further tested using 13 masses with a BIRAD score of 3 and the CAD results
are shown to agree with the radiologistâs classification after confirming in six
months follow up.
The present results demonstrate high sensitivity and specificity of the proposed
CAD system compared to prior research. The present research is therefore
intended to make contributions to the field by proposing a novel CAD system,
consists of series of well-selected image processing algorithms, to firstly classify
mass to benign or malignancy, secondly sub classify BIRAD 4 to three groups and
finally to interpret BIRAD 3 to BIRAD 2 without a need of follow up study
Pixel N-grams for Mammographic Image Classification
X-ray screening for breast cancer is an important public health initiative in the management of a leading cause of death for women. However, screening is expensive if mammograms are required to be manually assessed by radiologists. Moreover, manual screening is subject to perception and interpretation errors. Computer aided detection/diagnosis (CAD) systems can help radiologists as computer algorithms are good at performing image analysis consistently and repetitively. However, image features that enhance CAD classification accuracies are necessary for CAD systems to be deployed. Many CAD systems have been developed but the specificity and sensitivity is not high; in part because of challenges inherent in identifying effective features to be initially extracted from raw images. Existing feature extraction techniques can be grouped under three main approaches; statistical, spectral and structural. Statistical and spectral techniques provide global image features but often fail to distinguish between local pattern variations within an image. On the other hand, structural approach have given rise to the Bag-of-Visual-Words (BoVW) model, which captures local variations in an image, but typically do not consider spatial relationships between the visual âwordsâ. Moreover, statistical features and features based on BoVW models are computationally very expensive. Similarly, structural feature computation methods other than BoVW are also computationally expensive and strongly dependent upon algorithms that can segment an image to localize a region of interest likely to contain the tumour. Thus, classification algorithms using structural features require high resource computers. In order for a radiologist to classify the lesions on low resource computers such as Ipads, Tablets, and Mobile phones, in a remote location, it is necessary to develop computationally inexpensive classification algorithms. Therefore, the overarching aim of this research is to discover a feature extraction/image representation model which can be used to classify mammographic lesions with high accuracy, sensitivity and specificity along with low computational cost. For this purpose a novel feature extraction technique called âPixel N-gramsâ is proposed. The Pixel N-grams approach is inspired from the character N-gram concept in text categorization. Here, N number of consecutive pixel intensities are considered in a particular direction. The image is then represented with the help of histogram of occurrences of the Pixel N-grams in an image. Shape and texture of mammographic lesions play an important role in determining the malignancy of the lesion. It was hypothesized that the Pixel N-grams would be able to distinguish between various textures and shapes. Experiments carried out on benchmark texture databases and binary basic shapes database have demonstrated that the hypothesis was correct. Moreover, the Pixel N-grams were able to distinguish between various shapes irrespective of size and location of shape in an image. The efficacy of the Pixel N-gram technique was tested on mammographic database of primary digital mammograms sourced from a radiological facility in Australia (LakeImaging Pty Ltd) and secondary digital mammograms (benchmark miniMIAS database). A senior radiologist from LakeImaging provided real time de-identified high resolution mammogram images with annotated regions of interests (which were used as groundtruth), and valuable radiological diagnostic knowledge. Two types of classifications were observed on these two datasets. Normal/abnormal classification useful for automated screening and circumscribed/speculation/normal classification useful for automated diagnosis of breast cancer. The classification results on both the mammography datasets using Pixel N-grams were promising. Classification performance (Fscore, sensitivity and specificity) using Pixel N-gram technique was observed to be significantly better than the existing techniques such as intensity histogram, co-occurrence matrix based features and comparable with the BoVW features. Further, Pixel N-gram features are found to be computationally less complex than the co-occurrence matrix based features as well as BoVW features paving the way for mammogram classification on low resource computers. Although, the Pixel N-gram technique was designed for mammographic classification, it could be applied to other image classification applications such as diabetic retinopathy, histopathological image classification, lung tumour detection using CT images, brain tumour detection using MRI images, wound image classification and tooth decay classification using dentistry x-ray images. Further, texture and shape classification is also useful for classification of real world images outside the medical domain. Therefore, the pixel N-gram technique could be extended for applications such as classification of satellite imagery and other object detection tasks.Doctor of Philosoph