244 research outputs found
Recommended from our members
Saliency-driven system models for cell analysis with deep learning.
Background and objectivesSaliency refers to the visual perception quality that makes objects in a scene to stand out from others and attract attention. While computational saliency models can simulate the expert's visual attention, there is little evidence about how these models perform when used to predict the cytopathologist's eye fixations. Saliency models may be the key to instrumenting fast object detection on large Pap smear slides under real noisy conditions, artifacts, and cell occlusions. This paper describes how our computational schemes retrieve regions of interest (ROI) of clinical relevance using visual attention models. We also compare the performance of different computed saliency models as part of cell screening tasks, aiming to design a computer-aided diagnosis systems that supports cytopathologists.MethodWe record eye fixation maps from cytopathologists at work, and compare with 13 different saliency prediction algorithms, including deep learning. We develop cell-specific convolutional neural networks (CNN) to investigate the impact of bottom-up and top-down factors on saliency prediction from real routine exams. By combining the eye tracking data from pathologists with computed saliency models, we assess algorithms reliability in identifying clinically relevant cells.ResultsThe proposed cell-specific CNN model outperforms all other saliency prediction methods, particularly regarding the number of false positives. Our algorithm also detects the most clinically relevant cells, which are among the three top salient regions, with accuracy above 98% for all diseases, except carcinoma (87%). Bottom-up methods performed satisfactorily, with saliency maps that enabled ROI detection above 75% for carcinoma and 86% for other pathologies.ConclusionsROIs extraction using our saliency prediction methods enabled ranking the most relevant clinical areas within the image, a viable data reduction strategy to guide automatic analyses of Pap smear slides. Top-down factors for saliency prediction on cell images increases the accuracy of the estimated maps while bottom-up algorithms proved to be useful for predicting the cytopathologist's eye fixations depending on parameters, such as the number of false positive and negative. Our contributions are: comparison among 13 state-of-the-art saliency models to cytopathologists' visual attention and deliver a method that the associate the most conspicuous regions to clinically relevant cells
Custom Deep Learning Model for the Diagnosis of Cervical Carcinoma
Cancer is the second most common cause of death in the majority of the world due to late diagnosis. Most cancer cases are typically discovered at an advanced stage, which lowers the likelihood of recovery because proper therapy cannot be given at that time. In particular, for incurable cancers, which may result in a reduced life expectancy due to the rapid progression of the disease, the sooner cancer is identified, the more effective the therapy may be. Early detection also lessens the financial effects of cancer because treatment in the early stages is much cheaper than treatment in later stages.The method suggested is an end-to-end deep learning method in which the input photos are sent directly to the deep model, which makes the decision. The proposed Ensemble of deep learning modelIV3-DCNN to detect cancer in pap-test images. The model's precision, FScore, Specificity, Sensitivity, and accuracy of 99.4%, 99.23, 95.48, 97.9, and 99.2%. Last but not least, the suggested strategy would be very beneficial and successful, especially in low-income nations where referral mechanisms for patients with suspected cancer are frequently lacking, resulting in delayed and fragmented care
KLASIFIKASI TIPE SEL NORMAL/ABNORMAL BERDASARKAN CITRA PAP-SMEAR MENGGUNAKAN CONVOLUTIONAL NEURAL NETWORK
The classification of cell types plays an essential role in monitoring the growth of cancer cells. One of the methods to determine the cancer type is to analyze the pap-smear images manually. Nevertheless, the manual analysis of pap-smear images by the expert has several limitations, such as time-consuming and prone to misdiagnosis. For reducing the risks, it requires the automatic classification of cell types based on pap-smear images. This study utilizes the convolutional neural network (CNN) architectures to automatically classify the cell type into two-class categories (normal/abnormal) based on three features. These features, such as the local binary pattern, gray level co-occurrence matrix, and shape features, are extracted from pap-smear images. This study shows the performance of CNN achieved the maximum accuracy of 99.98%, 100.0%, 99.78% in training, validation, and testing data. Our approach also outperforms the performance of the baseline methods. Keywords : CNN, Classification, Cell, Neural Network, Pap-smea
Multi-class Cervical Cancer Classification using Transfer Learning-based Optimized SE-ResNet152 model in Pap Smear Whole Slide Images
Among the main factors contributing to death globally is cervical cancer, regardless of whether it can be avoided and treated if the afflicted tissues are removed early. Cervical screening programs must be made accessible to everyone and effectively, which is a difficult task that necessitates, among other things, identifying the population\u27s most vulnerable members. Therefore, we present an effective deep-learning method for classifying the multi-class cervical cancer disease using Pap smear images in this research. The transfer learning-based optimized SE-ResNet152 model is used for effective multi-class Pap smear image classification. The reliable significant image features are accurately extracted by the proposed network model. The network\u27s hyper-parameters are optimized using the Deer Hunting Optimization (DHO) algorithm. Five SIPaKMeD dataset categories and six CRIC dataset categories constitute the 11 classes for cervical cancer diseases. A Pap smear image dataset with 8838 images and various class distributions is used to evaluate the proposed method. The introduction of the cost-sensitive loss function throughout the classifier\u27s learning process rectifies the dataset\u27s imbalance. When compared to prior existing approaches on multi-class Pap smear image classification, 99.68% accuracy, 98.82% precision, 97.86% recall, and 98.64% F1-Score are achieved by the proposed method on the test set. For automated preliminary diagnosis of cervical cancer diseases, the proposed method produces better identification results in hospitals and cervical cancer clinics due to the positive classification results
- …