25 research outputs found

    Deep Semantic Segmentation and Multi-Class Skin Lesion Classification Based on Convolutional Neural Network

    Get PDF
    Skin cancer is developed due to abnormal cell growth. These cells are grown rapidly and destroy the normal skin cells. However, it's curable at an initial stage to reduce the patient's mortality rate. In this article, the method is proposed for localization, segmentation and classification of the skin lesion at an early stage. The proposed method contains three phases. In phase I, different types of the skin lesion are localized using tinyYOLOv2 model in which open neural network (ONNX) and squeeze Net model are used as a backbone. The features are extracted from depthconcat7 layer of squeeze Net and passed as an input to the tinyYOLOv2. The propose model accurately localize the affected part of the skin. In Phase II, 13-layer 3D-semantic segmentation model (01 input, 04 convolutional, 03 batch-normalization, 03 ReLU, softmax and pixel classification) is used for segmentation. In the proposed segmentation model, pixel classification layer is used for computing the overlap region between the segmented and ground truth images. Later in Phase III, extract deep features using ResNet-18 model and optimized features are selected using ant colony optimization (ACO) method. The optimized features vector is passed to the classifiers such as optimized (O)-SVM and O-NB. The proposed method is evaluated on the top MICCAI ISIC challenging 2017, 2018 and 2019 datasets. The proposed method accurately localized, segmented and classified the skin lesion at an early stage.Qatar University [IRCC-2020-009]

    An Integrated Design for Classification and Localization of Diabetic Foot Ulcer based on CNN and YOLOv2-DFU Models

    Get PDF
    Diabetes is a chronic disease, if not treated in time may lead to many complications including diabetic foot ulcers (DFU). DFU is a dangerous disease, it needs regular treatment otherwise it may lead towards foot amputation. The DFU is classified into two categories such as infection (bacteria) and the ischaemia (inadequate supply of the blood). The DFU detection at an initial phase is a tough procedure. Therefore in this research work 16 layers convolutional neural network (CNN) for example 01 input, 03 convolutional, 03 batch-normalization, 01 average pooling, 01 skips convolutional, 03 ReLU, 01 add (element-wise addition of two inputs), fully connected, softmax and classification output layers for classification and YOLOv2-DFU for localization of infection/ischaemia models are proposed. In the classification phase, deep features are extracted and supplied to the number of classifiers such as KNN, DT, Ensemble, softmax, and NB to analyze the classification results for the selection of best classifiers. After the experimentation, we observed that DT and softmax achieved consistent results for the detection of ischaemia/infection in all performance metrics such as sensitivity, specificity, and accuracy as compared with other classifiers. In addition, after the classification, the Gradient-weighted class activation mapping (Grad-Cam) model is used to visualize the high-level features of the infected region for better understanding. The classified images are passed to the YOLOv2-DFU network for infected region localization. The Shuffle network is utilized as a mainstay of the YOLOv2 model in which bottleneck extracted features through ReLU node-199 layer and passed to the YOLOv2 model. The proposed method is validated on the newly developed DFU-Part (B) dataset and the results are compared with the latest published work using the same dataset

    Recognition of different types of leukocytes using YOLoV2 and optimized bag-of-features

    Get PDF
    White blood cells (WBCs) protect human body against different types of infections including fungal, parasitic, viral, and bacterial. The detection of abnormal regions in WBCs is a difficult task. Therefore a method is proposed for the localization of WBCs based on YOLOv2-Nucleus-Cytoplasm, which contains darkNet-19 as a basenetwork of the YOLOv2 model. In this model features are extracted from LeakyReLU-18 of darkNet-19 and supplied as an input to the YOLOv2 model. The YOLOv2-Nucleus-Cytoplasm model localizes and classifies the WBCs with maximum score labels. It also localize the WBCs into the blast and non-blast cells. After localization, the bag-of-features are extracted and optimized by using particle swarm optimization(PSO). The improved feature vector is fed to classifiers i.e., optimized naĂŻve Bayes (O-NB) & optimized discriminant analysis (O-DA) for WBCs classification. The experiments are performed on LISC, ALL-IDB1, and ALL-IDB2 datasets

    Explainable Neural Network for Classification of Cotton Leaf Diseases

    No full text
    Every nation’s development depends on agriculture. The term “cash crops” refers to cotton and other important crops. Most pathogens that significantly harm crops also impact cotton. Numerous diseases that influence yield via the leaf, such as powdery mildew, leaf curl, leaf spot, target spot, bacterial blight, and nutrient deficiencies, can affect cotton. Early disease detection protects crops from additional harm. Computerized methods perform a vital role in cotton leaf disease detection at an early stage. The method consists of two core steps such as feature extraction and classification. First, in the proposed method, data augmentation is applied to balance the input data. After that, features are extracted from a pre-trained VGG-16 model and passed to 11 fully convolutional layers, which freeze the majority and randomly initialize convolutional features to subsequently generate a score of the anomaly map, which defines the probability of the lesion region. The proposed model is trained on the selected hyperparameters that produce great classification results. The proposed model performance is evaluated on two publicly available Kaggle datasets, Cotton Leaf and Disease. The proposed method provides 99.99% accuracy, which is competent compared to existing methods

    Clinically acquired new challenging dataset for brain SOL segmentation: AJBDS-2023

    No full text
    Space-occupying lesions (SOL) brain detected on brain MRI are benign and malignant tumors. Several brain tumor segmentation algorithms have been developed but there is a need for a clinically acquired dataset that is used for real-time images. This research is done to facilitate reporting of MRI done for brain tumor detection by incorporating computer-aided detection. Another objective was to make reporting unbiased by decreasing inter-observer errors and expediting daily reporting sessions to decrease radiologists’ workload. This is an experimental study. The proposed dataset contains clinically acquired multiplanar, multi-sequential MRI slices (MPMSI) which are used as input to the segmentation model without any preprocessing. The proposed AJBDS-2023 consists of 10667 images of real patients imaging data with a size of 320*320*3. Acquired images have T1W, TW2, Flair, T1W contrast, ADC, and DWI sequences. Pixel-based ground-truth annotated images of the tumor core and edema of 6334 slices are made manually under the supervision of a radiologist. Quantitative assessment of AJBDS-2023 images is done by a novel U-network on 4333 MRI slices. The diagnostic accuracy of our algorithm U-Net trained on AJBDS-2023 was 77.4 precision, 82.3 DSC, 87.4 specificity, 93.8 sensitivity, and 90.4 confidence interval. An experimental analysis of AJBDS-2023 done by the U-Net segmentation model proves that the proposed AJBDS-2023 dataset has images without preprocessing, which is more challenging and provides a more realistic platform for evaluation and analysis of newly developed algorithms in this domain and helps radiologists in MRI brain reporting more realistically

    Visual Geometry Group based on U-Shaped Model for Liver/Liver Tumor Segmentation

    No full text
    Liver cancer is the primary reason of death around the globe. Manually detecting the infected tissues is a challenging and time-consuming task. The computerized methods help make accurate decisions and therapy processes. The segmentation accuracy might be increased to reduce the loss rate. Semantic segmentation performs a vital role in infected liver region segmentation. This article proposes a method that consists of two major steps; first, the local Laplacian filter is applied to improve the image quality. The second is the proposed semantic segmentation model in which features are extracted to the pre-trained VGG16 model and passed to the U-shaped network. This model consists of 51 layers: input, 23 convolutional, 4 max pooling, 4 transpose convolutional, 4 concatenated, 8 activation, and 7 batch-normalization. The proposed segmentation framework is trained on the selected hyperparameters that reduce the loss rate and increase the segmentation accuracy. The proposed approach more precisely segments the infected liver region. The proposed approach performance is accessed on two datasets such as 3DIRCADB and LiTS17. The proposed framework provides an average dice score of 0.98, which is far better compared to the existing works

    DeepLabv3+-Based Segmentation and Best Features Selection Using Slime Mould Algorithm for Multi-Class Skin Lesion Classification

    No full text
    The development of abnormal cell growth is caused by different pathological alterations and some genetic disorders. This alteration in skin cells is very dangerous and life-threatening, and its timely identification is very essential for better treatment and safe cure. Therefore, in the present article, an approach is proposed for skin lesions’ segmentation and classification. So, in the proposed segmentation framework, pre-trained Mobilenetv2 is utilised in the act of the back pillar of the DeepLabv3+ model and trained on the optimum parameters that provide significant improvement for infected skin lesions’ segmentation. The multi-classification of the skin lesions is carried out through feature extraction from pre-trained DesneNet201 with N × 1000 dimension, out of which informative features are picked from the Slim Mould Algorithm (SMA) and input to SVM and KNN classifiers. The proposed method provided a mean ROC of 0.95 ± 0.03 on MED-Node, 0.97 ± 0.04 on PH2, 0.98 ± 0.02 on HAM-10000, and 0.97 ± 0.00 on ISIC-2019 datasets
    corecore