24 research outputs found
Skin Lesion Analyser: An Efficient Seven-Way Multi-Class Skin Cancer Classification Using MobileNet
Skin cancer, a major form of cancer, is a critical public health problem with
123,000 newly diagnosed melanoma cases and between 2 and 3 million non-melanoma
cases worldwide each year. The leading cause of skin cancer is high exposure of
skin cells to UV radiation, which can damage the DNA inside skin cells leading
to uncontrolled growth of skin cells. Skin cancer is primarily diagnosed
visually employing clinical screening, a biopsy, dermoscopic analysis, and
histopathological examination. It has been demonstrated that the dermoscopic
analysis in the hands of inexperienced dermatologists may cause a reduction in
diagnostic accuracy. Early detection and screening of skin cancer have the
potential to reduce mortality and morbidity. Previous studies have shown Deep
Learning ability to perform better than human experts in several visual
recognition tasks. In this paper, we propose an efficient seven-way automated
multi-class skin cancer classification system having performance comparable
with expert dermatologists. We used a pretrained MobileNet model to train over
HAM10000 dataset using transfer learning. The model classifies skin lesion
image with a categorical accuracy of 83.1 percent, top2 accuracy of 91.36
percent and top3 accuracy of 95.34 percent. The weighted average of precision,
recall, and f1-score were found to be 0.89, 0.83, and 0.83 respectively. The
model has been deployed as a web application for public use at
(https://saketchaturvedi.github.io). This fast, expansible method holds the
potential for substantial clinical impact, including broadening the scope of
primary care practice and augmenting clinical decision-making for dermatology
specialists.Comment: This is a pre-copyedited version of a contribution published in
Advances in Intelligent Systems and Computing, Hassanien A., Bhatnagar R.,
Darwish A. (eds) published by Chaturvedi S.S., Gupta K., Prasad P.S. The
definitive authentication version is available online via
https://doi.org/10.1007/978-981-15-3383-9_1
Multi-Resolution Fully Convolutional Neural Networks for Monaural Audio Source Separation
In deep neural networks with convolutional layers, each layer typically has
fixed-size/single-resolution receptive field (RF). Convolutional layers with a
large RF capture global information from the input features, while layers with
small RF size capture local details with high resolution from the input
features. In this work, we introduce novel deep multi-resolution fully
convolutional neural networks (MR-FCNN), where each layer has different RF
sizes to extract multi-resolution features that capture the global and local
details information from its input features. The proposed MR-FCNN is applied to
separate a target audio source from a mixture of many audio sources.
Experimental results show that using MR-FCNN improves the performance compared
to feedforward deep neural networks (DNNs) and single resolution deep fully
convolutional neural networks (FCNNs) on the audio source separation problem.Comment: arXiv admin note: text overlap with arXiv:1703.0801
A Comparative Analysis of Transfer Learning-based Techniques for the Classification of Melanocytic Nevi
Skin cancer is a fatal manifestation of cancer. Unrepaired deoxyribo-nucleic
acid (DNA) in skin cells, causes genetic defects in the skin and leads to skin
cancer. To deal with lethal mortality rates coupled with skyrocketing costs of
medical treatment, early diagnosis is mandatory. To tackle these challenges,
researchers have developed a variety of rapid detection tools for skin cancer.
Lesion-specific criteria are utilized to distinguish benign skin cancer from
malignant melanoma. In this study, a comparative analysis has been performed on
five Transfer Learning-based techniques that have the potential to be leveraged
for the classification of melanocytic nevi. These techniques are based on deep
convolutional neural networks (DCNNs) that have been pre-trained on thousands
of open-source images and are used for day-to-day classification tasks in many
instances.Comment: 12 pages, 5 figures, submitted to International Conference on
Advances and Applications of Artificial Intelligence and Machine Learning
(ICAAAIML) 2022, to be published in Springer's Lecture Notes in Electrical
Engineerin
Mobile learning architecture using fog computing and adaptive data streaming
With the huge development in mobile and network fields, sensor technologies and fog computing help the students for more effective learning, flexible and in and effective manner from anywhere. Using the mobile device for learn encourage the transition to mobile computing (cloud and fog computing) which is led to the ability to design customized system that help student to learn via context aware learning which can be done by set the user preference and use proper methods to show only related manner subject. The presented study works on developing a system of e-learning which has been on the basis of fog computing concepts with deep learning approaches utilized for classification to the data content for accomplishing the context aware learning and use the adaptation of video quality using special equation and the data encrypted and decrypted using 3DES algorithm to ensure the security side of the operation
A Multi-Channel Uncertainty-Aware Multi-Resolution Network for MR to CT Synthesis
Synthesising computed tomography (CT) images from magnetic resonance images (MRI) plays an important role in the field of medical image analysis, both for quantification and diagnostic purposes. Convolutional neural networks (CNNs) have achieved state-of-the-art results in image-to-image translation for brain applications. However, synthesising whole-body images remains largely uncharted territory, involving many challenges, including large image size and limited field of view, complex spatial context, and anatomical differences between images acquired at different times. We propose the use of an uncertainty-aware multi-channel multi-resolution 3D cascade network specifically aiming for whole-body MR to CT synthesis. The Mean Absolute Error on the synthetic CT generated with the MultiResunc network (73.90 HU) is compared to multiple baseline CNNs like 3D U-Net (92.89 HU), HighRes3DNet (89.05 HU) and deep boosted regression (77.58 HU) and shows superior synthesis performance. We ultimately exploit the extrapolation properties of the MultiRes networks on sub-regions of the body
An Ensemble of Statistical Metadata and CNN Classification of Class Imbalanced Skin Lesion Data
Skin Cancer is one of the most widely present forms of cancer. The correct classification of skin lesions as malignant or benign is a complex process that has to be undertaken by experienced specialists. Another major issue of the class imbalance of data causes a bias in the results of classification. This article presents a novel approach to the usage of metadata of skin lesions\u27 images to classify them. The usage of techniques addresses the problem of class imbalance to nullify the imbalances. Further, the use of a convolutional neural network (CNN) is proposed to fine-tune the skin lesion data classification. Ultimately, it is proven that an ensemble of statistical metadata analysis and CNN usage would result in the highest accuracy of skin color classification instead of using the two techniques separately
Hierarchical Action Classification with Network Pruning
Research on human action classification has made significant progresses in
the past few years. Most deep learning methods focus on improving performance
by adding more network components. We propose, however, to better utilize
auxiliary mechanisms, including hierarchical classification, network pruning,
and skeleton-based preprocessing, to boost the model robustness and
performance. We test the effectiveness of our method on four commonly used
testing datasets: NTU RGB+D 60, NTU RGB+D 120, Northwestern-UCLA Multiview
Action 3D, and UTD Multimodal Human Action Dataset. Our experiments show that
our method can achieve either comparable or better performance on all four
datasets. In particular, our method sets up a new baseline for NTU 120, the
largest dataset among the four. We also analyze our method with extensive
comparisons and ablation studies