24 research outputs found

    Skin Lesion Analyser: An Efficient Seven-Way Multi-Class Skin Cancer Classification Using MobileNet

    Full text link
    Skin cancer, a major form of cancer, is a critical public health problem with 123,000 newly diagnosed melanoma cases and between 2 and 3 million non-melanoma cases worldwide each year. The leading cause of skin cancer is high exposure of skin cells to UV radiation, which can damage the DNA inside skin cells leading to uncontrolled growth of skin cells. Skin cancer is primarily diagnosed visually employing clinical screening, a biopsy, dermoscopic analysis, and histopathological examination. It has been demonstrated that the dermoscopic analysis in the hands of inexperienced dermatologists may cause a reduction in diagnostic accuracy. Early detection and screening of skin cancer have the potential to reduce mortality and morbidity. Previous studies have shown Deep Learning ability to perform better than human experts in several visual recognition tasks. In this paper, we propose an efficient seven-way automated multi-class skin cancer classification system having performance comparable with expert dermatologists. We used a pretrained MobileNet model to train over HAM10000 dataset using transfer learning. The model classifies skin lesion image with a categorical accuracy of 83.1 percent, top2 accuracy of 91.36 percent and top3 accuracy of 95.34 percent. The weighted average of precision, recall, and f1-score were found to be 0.89, 0.83, and 0.83 respectively. The model has been deployed as a web application for public use at (https://saketchaturvedi.github.io). This fast, expansible method holds the potential for substantial clinical impact, including broadening the scope of primary care practice and augmenting clinical decision-making for dermatology specialists.Comment: This is a pre-copyedited version of a contribution published in Advances in Intelligent Systems and Computing, Hassanien A., Bhatnagar R., Darwish A. (eds) published by Chaturvedi S.S., Gupta K., Prasad P.S. The definitive authentication version is available online via https://doi.org/10.1007/978-981-15-3383-9_1

    Multi-Resolution Fully Convolutional Neural Networks for Monaural Audio Source Separation

    Get PDF
    In deep neural networks with convolutional layers, each layer typically has fixed-size/single-resolution receptive field (RF). Convolutional layers with a large RF capture global information from the input features, while layers with small RF size capture local details with high resolution from the input features. In this work, we introduce novel deep multi-resolution fully convolutional neural networks (MR-FCNN), where each layer has different RF sizes to extract multi-resolution features that capture the global and local details information from its input features. The proposed MR-FCNN is applied to separate a target audio source from a mixture of many audio sources. Experimental results show that using MR-FCNN improves the performance compared to feedforward deep neural networks (DNNs) and single resolution deep fully convolutional neural networks (FCNNs) on the audio source separation problem.Comment: arXiv admin note: text overlap with arXiv:1703.0801

    A Comparative Analysis of Transfer Learning-based Techniques for the Classification of Melanocytic Nevi

    Full text link
    Skin cancer is a fatal manifestation of cancer. Unrepaired deoxyribo-nucleic acid (DNA) in skin cells, causes genetic defects in the skin and leads to skin cancer. To deal with lethal mortality rates coupled with skyrocketing costs of medical treatment, early diagnosis is mandatory. To tackle these challenges, researchers have developed a variety of rapid detection tools for skin cancer. Lesion-specific criteria are utilized to distinguish benign skin cancer from malignant melanoma. In this study, a comparative analysis has been performed on five Transfer Learning-based techniques that have the potential to be leveraged for the classification of melanocytic nevi. These techniques are based on deep convolutional neural networks (DCNNs) that have been pre-trained on thousands of open-source images and are used for day-to-day classification tasks in many instances.Comment: 12 pages, 5 figures, submitted to International Conference on Advances and Applications of Artificial Intelligence and Machine Learning (ICAAAIML) 2022, to be published in Springer's Lecture Notes in Electrical Engineerin

    Mobile learning architecture using fog computing and adaptive data streaming

    Get PDF
    With the huge development in mobile and network fields, sensor technologies and fog computing help the students for more effective learning, flexible and in and effective manner from anywhere. Using the mobile device for learn encourage the transition to mobile computing (cloud and fog computing) which is led to the ability to design customized system that help student to learn via context aware learning which can be done by set the user preference and use proper methods to show only related manner subject. The presented study works on developing a system of e-learning which has been on the basis of fog computing concepts with deep learning approaches utilized for classification to the data content for accomplishing the context aware learning and use the adaptation of video quality using special equation and the data encrypted and decrypted using 3DES algorithm to ensure the security side of the operation

    A Multi-Channel Uncertainty-Aware Multi-Resolution Network for MR to CT Synthesis

    Get PDF
    Synthesising computed tomography (CT) images from magnetic resonance images (MRI) plays an important role in the field of medical image analysis, both for quantification and diagnostic purposes. Convolutional neural networks (CNNs) have achieved state-of-the-art results in image-to-image translation for brain applications. However, synthesising whole-body images remains largely uncharted territory, involving many challenges, including large image size and limited field of view, complex spatial context, and anatomical differences between images acquired at different times. We propose the use of an uncertainty-aware multi-channel multi-resolution 3D cascade network specifically aiming for whole-body MR to CT synthesis. The Mean Absolute Error on the synthetic CT generated with the MultiResunc network (73.90 HU) is compared to multiple baseline CNNs like 3D U-Net (92.89 HU), HighRes3DNet (89.05 HU) and deep boosted regression (77.58 HU) and shows superior synthesis performance. We ultimately exploit the extrapolation properties of the MultiRes networks on sub-regions of the body

    An Ensemble of Statistical Metadata and CNN Classification of Class Imbalanced Skin Lesion Data

    Get PDF
    Skin Cancer is one of the most widely present forms of cancer. The correct classification of skin lesions as malignant or benign is a complex process that has to be undertaken by experienced specialists. Another major issue of the class imbalance of data causes a bias in the results of classification. This article presents a novel approach to the usage of metadata of skin lesions\u27 images to classify them. The usage of techniques addresses the problem of class imbalance to nullify the imbalances. Further, the use of a convolutional neural network (CNN) is proposed to fine-tune the skin lesion data classification. Ultimately, it is proven that an ensemble of statistical metadata analysis and CNN usage would result in the highest accuracy of skin color classification instead of using the two techniques separately

    Hierarchical Action Classification with Network Pruning

    Full text link
    Research on human action classification has made significant progresses in the past few years. Most deep learning methods focus on improving performance by adding more network components. We propose, however, to better utilize auxiliary mechanisms, including hierarchical classification, network pruning, and skeleton-based preprocessing, to boost the model robustness and performance. We test the effectiveness of our method on four commonly used testing datasets: NTU RGB+D 60, NTU RGB+D 120, Northwestern-UCLA Multiview Action 3D, and UTD Multimodal Human Action Dataset. Our experiments show that our method can achieve either comparable or better performance on all four datasets. In particular, our method sets up a new baseline for NTU 120, the largest dataset among the four. We also analyze our method with extensive comparisons and ablation studies
    corecore