1,598 research outputs found

    Inter-comparison of medical image segmentation algorithms

    Get PDF
    Segmentation of images is a vital part of medical image processing, and MRI (Magnetic Resonance Imaging) is already recognized as a very important tool for clinical diagnosis. In this thesis, comparisons between different segmentation algorithms are carried out, specifically on brain MRI images. Initial parts of the thesis provide the background to the project, and an introduction to the basic principles of MRI, respectively, followed by parameter definitions and MRI image artifacts. The next part briefly covers various image pre-processing techniques which are required, and this is followed with a review of the major segmentation techniques which are available, including thresholding, region growing, clustering, and K-Means clustering. The concept of fuzzy logic is also introduced here, and the chapter concludes with a discussion of fuzzy logic based segmentation algorithms such as Fuzzy C-Means (FCM) and Improved Fuzzy C-Means (IFCM) clustering algorithms. The following part provides details concerning the source, type and parameters of the data (images) used for this thesis. Evaluation and inter-comparisons between a number of different segmentation algorithms are given in near concluding part, finally, conclusions and suggestions for future research are provided in last part. Qualitative comparisons on real images and quantitative comparisons on simulated images were performed. Both qualitative and quantitative comparisons demonstrated that fuzzy logic based segmentation algorithms are superior in comparison with classical segmentation algorithms. Edge-based segmentation algorithms demonstrated the poorest performance of all; K-means and IFCM clustering algorithms performed better, and FCM demonstrated the best performance of all. However, it should be noted that IFCM was not properly evaluated due to time restrictions in code generation, testing and evaluation.Segmentation of images is a vital part of medical image processing, and MRI (Magnetic Resonance Imaging) is already recognized as a very important tool for clinical diagnosis. In this thesis, comparisons between different segmentation algorithms are carried out, specifically on brain MRI images. Initial parts of the thesis provide the background to the project, and an introduction to the basic principles of MRI, respectively, followed by parameter definitions and MRI image artifacts. The next part briefly covers various image pre-processing techniques which are required, and this is followed with a review of the major segmentation techniques which are available, including thresholding, region growing, clustering, and K-Means clustering. The concept of fuzzy logic is also introduced here, and the chapter concludes with a discussion of fuzzy logic based segmentation algorithms such as Fuzzy C-Means (FCM) and Improved Fuzzy C-Means (IFCM) clustering algorithms. The following part provides details concerning the source, type and parameters of the data (images) used for this thesis. Evaluation and inter-comparisons between a number of different segmentation algorithms are given in near concluding part, finally, conclusions and suggestions for future research are provided in last part. Qualitative comparisons on real images and quantitative comparisons on simulated images were performed. Both qualitative and quantitative comparisons demonstrated that fuzzy logic based segmentation algorithms are superior in comparison with classical segmentation algorithms. Edge-based segmentation algorithms demonstrated the poorest performance of all; K-means and IFCM clustering algorithms performed better, and FCM demonstrated the best performance of all. However, it should be noted that IFCM was not properly evaluated due to time restrictions in code generation, testing and evaluation

    An Intelligent Hybrid Optimization with Deep Learning model-based Schizophrenia Identification from Structural MRI

    Get PDF
    One of the fatal diseases that claim women while they are pregnant or nursing is schizophrenia. Despite several developments and symptoms, it can be challenging to discern between benign and malignant conditions. The main and most popular imaging method to predict Schizophrenia is MR Images. Furthermore, a few earlier models had a definite accuracy when diagnosing the condition. Stable MRI criteria must also be implemented immediately. Compared to other imaging technologies, the MRI imaging method is the simplest, safest, and most common for predicting Schizophrenia. The following factors are mostly involved in the subprocess for the initial MRI image. Before calculating the length between the sample point and the cluster center, the initial cluster center of the sample is identified. Classification is done according to how far the sample point is from the cluster center. The picture is then generated once the new cluster center has been derived using the classification history and verified to match the cluster convergence condition. A grey wolf optimization-based convolutional neural network approach is offered to get beyond the limitations and find schizophrenia, whether its hazardous or not. Many MRI images or datasets are analyzed in a short time, and the results show a more accurate or higher rate of schizophrenia recognition

    A Fully Automatic Segmentation Method for Breast Ultrasound Images

    Get PDF
    Breast cancer is the second leading cause of death of women worldwide. Accurate lesion boundary detection is important for breast cancer diagnosis. Since many crucial features for discriminating benign and malignant lesions are based on the contour, shape, and texture of the lesion, an accurate segmentation method is essential for a successful diagnosis. Ultrasound is an effective screening tool and primarily useful for differentiating benign and malignant lesions. However, due to inherent speckle noise and low contrast of breast ultrasound imaging, automatic lesion segmentation is still a challenging task. This research focuses on developing a novel, effective, and fully automatic lesion segmentation method for breast ultrasound images. By incorporating empirical domain knowledge of breast structure, a region of interest is generated. Then, a novel enhancement algorithm (using a novel phase feature) and a newly developed neutrosophic clustering method are developed to detect the precise lesion boundary. Neutrosophy is a recently introduced branch of philosophy that deals with paradoxes, contradictions, antitheses, and antinomies. When neutrosophy is used to segment images with vague boundaries, its unique ability to deal with uncertainty is brought to bear. In this work, we apply neutrosophy to breast ultrasound image segmentation and propose a new clustering method named neutrosophic l-means. We compare the proposed method with traditional fuzzy c-means clustering and three other well-developed segmentation methods for breast ultrasound images, using the same database. Both accuracy and time complexity are analyzed. The proposed method achieves the best accuracy (TP rate is 94.36%, FP rate is 8.08%, and similarity rate is 87.39%) with a fairly rapid processing speed (about 20 seconds). Sensitivity analysis shows the robustness of the proposed method as well. Cases with multiple-lesions and severe shadowing effect (shadow areas having similar intensity values of the lesion and tightly connected with the lesion) are not included in this study

    On the Stability of Region Count in the Parameter Space of Image Analysis Methods

    Get PDF
    In this dissertation a novel bottom-up computer vision approach is proposed. This approach is based upon quantifying the stability of the number of regions or count in a multi-dimensional parameter scale-space. The stability analysis comes from the properties of flat areas in the region count space generated through bottom-up algorithms of thresholding and region growing, hysteresis thresholding, variance-based region growing. The parameters used can be threshold, region growth, intensity statistics and other low-level parameters. The advantages and disadvantages of top-down, bottom-up and hybrid computational models are discussed. The approaches of scale-space, perceptual organization and clustering methods in computer vision are also analyzed, and the difference between our approach and these approaches is clarified. An overview of our stable count idea and implementation of three algorithms derived from this idea are presented. The algorithms are applied to real-world images as well as simulated signals. We have developed three experiments based upon our framework of stable region count. The experiments are using flower detector, peak detector and retinal image lesion detector respectively to process images and signals. The results from these experiments all suggest that our computer vision framework can solve different image and signal problems and provide satisfactory solutions. In the end future research directions and improvements are proposed

    Video enhancement : content classification and model selection

    Get PDF
    The purpose of video enhancement is to improve the subjective picture quality. The field of video enhancement includes a broad category of research topics, such as removing noise in the video, highlighting some specified features and improving the appearance or visibility of the video content. The common difficulty in this field is how to make images or videos more beautiful, or subjectively better. Traditional approaches involve lots of iterations between subjective assessment experiments and redesigns of algorithm improvements, which are very time consuming. Researchers have attempted to design a video quality metric to replace the subjective assessment, but so far it is not successful. As a way to avoid heuristics in the enhancement algorithm design, least mean square methods have received considerable attention. They can optimize filter coefficients automatically by minimizing the difference between processed videos and desired versions through a training. However, these methods are only optimal on average but not locally. To solve the problem, one can apply the least mean square optimization for individual categories that are classified by local image content. The most interesting example is Kondo’s concept of local content adaptivity for image interpolation, which we found could be generalized into an ideal framework for content adaptive video processing. We identify two parts in the concept, content classification and adaptive processing. By exploring new classifiers for the content classification and new models for the adaptive processing, we have generalized a framework for more enhancement applications. For the part of content classification, new classifiers have been proposed to classify different image degradations such as coding artifacts and focal blur. For the coding artifact, a novel classifier has been proposed based on the combination of local structure and contrast, which does not require coding block grid detection. For the focal blur, we have proposed a novel local blur estimation method based on edges, which does not require edge orientation detection and shows more robust blur estimation. With these classifiers, the proposed framework has been extended to coding artifact robust enhancement and blur dependant enhancement. With the content adaptivity to more image features, the number of content classes can increase significantly. We show that it is possible to reduce the number of classes without sacrificing much performance. For the part of model selection, we have introduced several nonlinear filters to the proposed framework. We have also proposed a new type of nonlinear filter, trained bilateral filter, which combines both advantages of the original bilateral filter and the least mean square optimization. With these nonlinear filters, the proposed framework show better performance than with linear filters. Furthermore, we have shown a proof-of-concept for a trained approach to obtain contrast enhancement by a supervised learning. The transfer curves are optimized based on the classification of global or local image content. It showed that it is possible to obtain the desired effect by learning from other computationally expensive enhancement algorithms or expert-tuned examples through the trained approach. Looking back, the thesis reveals a single versatile framework for video enhancement applications. It widens the application scope by including new content classifiers and new processing models and offers scalabilities with solutions to reduce the number of classes, which can greatly accelerate the algorithm design

    Monocular Camera Viewpoint-Invariant Vehicular Traffic Segmentation and Classification Utilizing Small Datasets

    Get PDF
    The work presented here develops a computer vision framework that is view angle independent for vehicle segmentation and classification from roadway traffic systems installed by the Virginia Department of Transportation (VDOT). An automated technique for extracting a region of interest is discussed to speed up the processing. The VDOT traffic videos are analyzed for vehicle segmentation using an improved robust low-rank matrix decomposition technique. It presents a new and effective thresholding method that improves segmentation accuracy and simultaneously speeds up the segmentation processing. Size and shape physical descriptors from morphological properties and textural features from the Histogram of Oriented Gradients (HOG) are extracted from the segmented traffic. Furthermore, a multi-class support vector machine classifier is employed to categorize different traffic vehicle types, including passenger cars, passenger trucks, motorcycles, buses, and small and large utility trucks. It handles multiple vehicle detections through an iterative k-means clustering over-segmentation process. The proposed algorithm reduced the processed data by an average of 40%. Compared to recent techniques, it showed an average improvement of 15% in segmentation accuracy, and it is 55% faster than the compared segmentation techniques on average. Moreover, a comparative analysis of 23 different deep learning architectures is presented. The resulting algorithm outperformed the compared deep learning algorithms for the quality of vehicle classification accuracy. Furthermore, the timing analysis showed that it could operate in real-time scenarios

    Adaptive threshold optimisation for colour-based lip segmentation in automatic lip-reading systems

    Get PDF
    A thesis submitted to the Faculty of Engineering and the Built Environment, University of the Witwatersrand, Johannesburg, in ful lment of the requirements for the degree of Doctor of Philosophy. Johannesburg, September 2016Having survived the ordeal of a laryngectomy, the patient must come to terms with the resulting loss of speech. With recent advances in portable computing power, automatic lip-reading (ALR) may become a viable approach to voice restoration. This thesis addresses the image processing aspect of ALR, and focuses three contributions to colour-based lip segmentation. The rst contribution concerns the colour transform to enhance the contrast between the lips and skin. This thesis presents the most comprehensive study to date by measuring the overlap between lip and skin histograms for 33 di erent colour transforms. The hue component of HSV obtains the lowest overlap of 6:15%, and results show that selecting the correct transform can increase the segmentation accuracy by up to three times. The second contribution is the development of a new lip segmentation algorithm that utilises the best colour transforms from the comparative study. The algorithm is tested on 895 images and achieves percentage overlap (OL) of 92:23% and segmentation error (SE) of 7:39 %. The third contribution focuses on the impact of the histogram threshold on the segmentation accuracy, and introduces a novel technique called Adaptive Threshold Optimisation (ATO) to select a better threshold value. The rst stage of ATO incorporates -SVR to train the lip shape model. ATO then uses feedback of shape information to validate and optimise the threshold. After applying ATO, the SE decreases from 7:65% to 6:50%, corresponding to an absolute improvement of 1:15 pp or relative improvement of 15:1%. While this thesis concerns lip segmentation in particular, ATO is a threshold selection technique that can be used in various segmentation applications.MT201
    • …
    corecore