13 research outputs found
Hybrid CNN-SVD Based Prominent Feature Extraction and Selection for Grading Diabetic Retinopathy Using Extreme Learning Machine Algorithm
Funding Information: This work was supported in part by the National Research Foundation of Korea-Grant funded by the Government of Korea (Ministry of Science and ICT) under Grant NRF-2020R1A2B5B02002478, and in part by Sejong University through the Faculty Research Program under Grant 20212023Peer reviewedPublisher PD
An ensemble approach for the prediction of diabetes mellitus using a soft voting classifier with an explainable AI
Diabetes is a chronic disease that continues to be a primary and worldwide health concern since the health of the entire population has been affected by it. Over the years, many academics have attempted to develop a reliable diabetes prediction model using machine learning (ML) algorithms. However, these research investigations have had a minimal impact on clinical practice as the current studies focus mainly on improving the performance of complicated ML models while ignoring their explainability to clinical situations. Therefore, the physicians find it difficult to understand these models and rarely trust them for clinical use. In this study, a carefully constructed, efficient, and interpretable diabetes detection method using an explainable AI has been proposed. The Pima Indian diabetes dataset was used, containing a total of 768 instances where 268 are diabetic, and 500 cases are non-diabetic with several diabetic attributes. Here, six machine learning algorithms (artificial neural network (ANN), random forest (RF), support vector machine (SVM), logistic regression (LR), AdaBoost, XGBoost) have been used along with an ensemble classifier to diagnose the diabetes disease. For each machine learning model, global and local explanations have been produced using the Shapley additive explanations (SHAP), which are represented in different types of graphs to help physicians in understanding the model predictions. The balanced accuracy of the developed weighted ensemble model was 90% with a F1 score of 89% using a five-fold cross-validation (CV). The median values were used for the imputation of the missing values and the synthetic minority oversampling technique (SMOTETomek) was used to balance the classes of the dataset. The proposed approach can improve the clinical understanding of a diabetes diagnosis and help in taking necessary action at the very early stages of the disease
Enhancing agriculture through real-time grape leaf disease classification via an edge device with a lightweight CNN architecture and Grad-CAM
Crop diseases can significantly affect various aspects of crop cultivation, including crop yield, quality, production costs, and crop loss. The utilization of modern technologies such as image analysis via machine learning techniques enables early and precise detection of crop diseases, hence empowering farmers to effectively manage and avoid the occurrence of crop diseases. The proposed methodology involves the use of modified MobileNetV3Large model deployed on edge device for real-time monitoring of grape leaf disease while reducing computational memory demands and ensuring satisfactory classification performance. To enhance applicability of MobileNetV3Large, custom layers consisting of two dense layers were added, each followed by a dropout layer, helped mitigate overfitting and ensured that the model remains efficient. Comparisons among other models showed that the proposed model outperformed those with an average train and test accuracy of 99.66% and 99.42%, with a precision, recall, and F1 score of approximately 99.42%. The model was deployed on an edge device (Nvidia Jetson Nano) using a custom developed GUI app and predicted from both saved and real-time data with high confidence values. Grad-CAM visualization was used to identify and represent image areas that affect the convolutional neural network (CNN) classification decision-making process with high accuracy. This research contributes to the development of plant disease classification technologies for edge devices, which have the potential to enhance the ability of autonomous farming for farmers, agronomists, and researchers to monitor and mitigate plant diseases efficiently and effectively, with a positive impact on global food security
Diabetic Retinopathy Identification Using Parallel Convolutional Neural Network Based Feature Extractor and ELM Classifier
Diabetic retinopathy (DR) is an incurable retinal condition caused by excessive blood sugar that, if left untreated, can result in even blindness. A novel automated technique for DR detection has been proposed in this paper. To accentuate the lesions, the fundus images (FIs) were preprocessed using Contrast Limited Adaptive Histogram Equalization (CLAHE). A parallel convolutional neural network (PCNN) was employed for feature extraction and then the extreme learning machine (ELM) technique was utilized for the DR classification. In comparison to the similar CNN structure, the PCNN design uses fewer parameters and layers, which minimizes the time required to extract distinctive features. The effectiveness of the technique was evaluated on two datasets (Kaggle DR 2015 competition (Dataset 1; 34,984 FIs) and APTOS 2019 (3,662 FIs)), and the results are promising. For the two datasets mentioned, the proposed technique attained accuracies of 91.78 % and 97.27 % respectively. However, one of the study's subsidiary discoveries was that the proposed framework demonstrated stability for both larger and smaller datasets, as well as for balanced and imbalanced datasets. Furthermore, in terms of classifier performance metrics, model parameters and layers, and prediction time, the suggested approach outscored existing state-of-the-art models, which would add significant benefit for the medical practitioners in accurately identifying the DR
Explainable Transformer-Based Deep Learning Model for the Detection of Malaria Parasites from Blood Cell Images
Malaria is a life-threatening disease caused by female anopheles mosquito bites. Various plasmodium parasites spread in the victim’s blood cells and keep their life in a critical situation. If not treated at the early stage, malaria can cause even death. Microscopy is a familiar process for diagnosing malaria, collecting the victim’s blood samples, and counting the parasite and red blood cells. However, the microscopy process is time-consuming and can produce an erroneous result in some cases. With the recent success of machine learning and deep learning in medical diagnosis, it is quite possible to minimize diagnosis costs and improve overall detection accuracy compared with the traditional microscopy method. This paper proposes a multiheaded attention-based transformer model to diagnose the malaria parasite from blood cell images. To demonstrate the effectiveness of the proposed model, the gradient-weighted class activation map (Grad-CAM) technique was implemented to identify which parts of an image the proposed model paid much more attention to compared with the remaining parts by generating a heatmap image. The proposed model achieved a testing accuracy, precision, recall, f1-score, and AUC score of 96.41%, 96.99%, 95.88%, 96.44%, and 99.11%, respectively, for the original malaria parasite dataset and 99.25%, 99.08%, 99.42%, 99.25%, and 99.99%, respectively, for the modified dataset. Various hyperparameters were also finetuned to obtain optimum results, which were also compared with state-of-the-art (SOTA) methods for malaria parasite detection, and the proposed method outperformed the existing methods
Detection of various lung diseases including COVID-19 using extreme learning machine algorithm based on the features extracted from a lightweight CNN architecture
Around the world, several lung diseases such as pneumonia, cardiomegaly, and tuberculosis (TB) contribute to severe illness, hospitalization or even death, particularly for elderly and medically vulnerable patients. In the last few decades, several new types of lung-related diseases have taken the lives of millions of people, and COVID-19 has taken almost 6.27 million lives. To fight against lung diseases, timely and correct diagnosis with appropriate treatment is crucial in the current COVID-19 pandemic. In this study, an intelligent recognition system for seven lung diseases has been proposed based on machine learning (ML) techniques to aid the medical experts. Chest X-ray (CXR) images of lung diseases were collected from several publicly available databases. A lightweight convolutional neural network (CNN) has been used to extract characteristic features from the raw pixel values of the CXR images. The best feature subset has been identified using the Pearson Correlation Coefficient (PCC). Finally, the extreme learning machine (ELM) has been used to perform the classification task to assist faster learning and reduced computational complexity. The proposed CNN-PCC-ELM model achieved an accuracy of 96.22% with an Area Under Curve (AUC) of 99.48% for eight class classification. The outcomes from the proposed model demonstrated better performance than the existing state-of-the-art (SOTA) models in the case of COVID-19, pneumonia, and tuberculosis detection in both binary and multiclass classifications. For eight class classification, the proposed model achieved precision, recall and fi-score and ROC are 100%, 99%, 100% and 99.99% respectively for COVID-19 detection demonstrating its robustness. Therefore, the proposed model has overshadowed the existing pioneering models to accurately differentiate COVID-19 from the other lung diseases that can assist the medical physicians in treating the patient effectively
Fast and accurate fault detection and classification in transmission lines using extreme learning machine
To provide stability and a continuous supply of power, the detection and classification of faults in the transmission lines (TLs) are crucial in this modern age. It is required to remove a faulty section from a healthy section to provide safety and to minimize power loss due to the fault. In the contemporary world, machine learning (ML) is extensively used in every aspect of life. In this study, a spontaneous fault detection (FD) and fault classification (FC) system based on ML has been proposed. MATLAB Simulink was employed to simulate two different TLs and to generate normal and fault data (Per unit voltage and current) of ten different types. TL-1 consisted of a single generator and a single load whereas TL-2 consisted of two generators and three loads. Upon normalizing the data, an extreme learning machine (ELM) algorithm was used as the classifier. Two different ELM models were developed for FD and FC purposes through training. The method achieved fault classification accuracies of 99.18% and 99.09% for the TL-1 and TL-2 respectively. On the other hand, fault detection accuracies of 99.53% and 99.60% were achieved for the TL-1 and TL-2. The proposed ELM model compared to a traditional artificial neural network (ANN) model demonstrated relatively a shorter processing time and reduced computational complexity. In addition, the proposed method outperformed the existing state-of-the-art methods
Applying supervised contrastive learning for the detection of diabetic retinopathy and its severity levels from fundus images
Diabetic Retinopathy (DR) is a major complication in human eyes among the diabetic patients. Early detection of the DR can save many patients from permanent blindness. Various artificial intelligent based systems have been proposed and they outperform human analysis in accurate detection of the DR. In most of the traditional deep learning models, the cross-entropy is used as a common loss function in a single stage end-to-end training method. However, it has been recently identified that this loss function has some limitations such as poor margin leading to false results, sensitive to noisy data and hyperparameter variations. To overcome these issues, supervised contrastive learning (SCL) has been introduced. In this study, SCL method, a two-stage training method with supervised contrastive loss function was proposed for the first time to the best of authors' knowledge to identify the DR and its severity stages from fundus images (FIs) using “APTOS 2019 Blindness Detection” dataset. “Messidor-2” dataset was also used to conduct experiments for further validating the model's performance. Contrast Limited Adaptive Histogram Equalization (CLAHE) was applied for enhancing the image quality and the pre-trained Xception CNN model was deployed as the encoder with transfer learning. To interpret the SCL of the model, t-SNE method was used to visualize the embedding space (unit hyper sphere) composed of 128 D space into a 2 D space. The proposed model achieved a test accuracy of 98.36%, and AUC score of 98.50% to identify the DR (Binary classification) and a test accuracy of 84.364%, and AUC score of 93.819% for five stages grading with the APTOS 2019 dataset. Other evaluation metrics (precision, recall, F1-score) were also determined with APTOS 2019 as well as with Messidor-2 for analyzing the performance of the proposed model. It was also concluded that the proposed method achieved better performance in detecting the DR compared to the conventional CNN without SCL and other state-of-the-art methods
A Novel Method for Multivariant Pneumonia Classification based on Hybrid CNN-PCA Based Feature Extraction using Extreme Learning Machine with Chest X-Ray Images
In this era of COVID19, proper diagnosis and treatment for pneumonia are very important. Chest X-Ray (CXR) image analysis plays a vital role in the reliable diagnosis of pneumonia. An experienced radiologist is required for this. However, even for an experienced radiographer, it is quite difficult and timeconsuming to diagnose due to the fuzziness of CXR images. Also, identification can be erroneous due to the involvement of human judgment. Hence, an authentic and automated system can play an important role here. In this era of cutting-edge technology, deep learning (DL) is highly used in every sector. There are several existing methods to diagnose pneumonia but they have accuracy problems. In this study, an automatic pneumonia detection system has been proposed by applying the extreme learning machine (ELM) on the Kaggle CXR images (Pneumonia). Three models have been studied: classification using extreme learning machine (ELM), ELM with a hybrid convolutional neural network - principle component analysis (CNN-PCA) based feature extraction (ECP), and ECP with the CXR images which are contrast-enhanced by contrast limited adaptive histogram equalization (CLAHE). Among these three proposed methods, the final model provides an optimistic result. It achieves the recall score of 98% and accuracy score of 98.32% for multiclass pneumonia classification. On the other hand, a binary classification achieves 100% recall and 99.83% accuracy. The proposed method also outperforms the existing methods. The outcome has been compared using several benchmarks that include accuracy, precision, recall, etc
Parallel CNN-ELM: A Multiclass Classification of Chest X-Ray Images to Identify Seventeen Lung Diseases Including COVID-19
Numerous epidemic lung diseases such as COVID-19, tuberculosis (TB), and pneumonia have spread over the world, killing millions of people. Medical specialists have experienced challenges in correctly identifying these diseases due to their subtle differences in Chest X-ray images (CXR). To assist the medical experts, this study proposed a computer-aided lung illness identification method based on the CXR images. For the first time, 17 different forms of lung disorders were considered and the study was divided into six trials with each containing two, two, three, four, fourteen, and seventeen different forms of lung disorders. The proposed framework combined robust feature extraction capabilities of a lightweight parallel convolutional neural network (CNN) with the classification abilities of the extreme learning machine algorithm named CNN-ELM. An optimistic accuracy of 90.92% and an area under the curve (AUC) of 96.93% was achieved when 17 classes were classified side by side. It also accurately identified COVID-19 and TB with 99.37% and 99.98% accuracy, respectively, in 0.996 microseconds for a single image. Additionally, the current results also demonstrated that the framework could outperform the existing state-of-the-art (SOTA) models. On top of that, a secondary conclusion drawn from this study was that the prospective framework retained its effectiveness over a range of real-world environments, including balanced-unbalanced or large-small datasets, large multiclass or simple binary class, and high- or low-resolution images. A prototype Android App was also developed to establish the potential of the framework in real-life implementation