217 research outputs found
Cats or CAT scans: transfer learning from natural or medical image source datasets?
Transfer learning is a widely used strategy in medical image analysis.
Instead of only training a network with a limited amount of data from the
target task of interest, we can first train the network with other, potentially
larger source datasets, creating a more robust model. The source datasets do
not have to be related to the target task. For a classification task in lung CT
images, we could use both head CT images, or images of cats, as the source.
While head CT images appear more similar to lung CT images, the number and
diversity of cat images might lead to a better model overall. In this survey we
review a number of papers that have performed similar comparisons. Although the
answer to which strategy is best seems to be "it depends", we discuss a number
of research directions we need to take as a community, to gain more
understanding of this topic.Comment: Accepted to Current Opinion in Biomedical Engineerin
Deep Learning based HEp-2 Image Classification: A Comprehensive Review
Classification of HEp-2 cell patterns plays a significant role in the
indirect immunofluorescence test for identifying autoimmune diseases in the
human body. Many automatic HEp-2 cell classification methods have been proposed
in recent years, amongst which deep learning based methods have shown
impressive performance. This paper provides a comprehensive review of the
existing deep learning based HEp-2 cell image classification methods. These
methods perform HEp-2 image classification at two levels, namely, cell-level
and specimen-level. Both levels are covered in this review. At each level, the
methods are organized with a deep network usage based taxonomy. The core idea,
notable achievements, and key strengths and weaknesses of each method are
critically analyzed. Furthermore, a concise review of the existing HEp-2
datasets that are commonly used in the literature is given. The paper ends with
a discussion on novel opportunities and future research directions in this
field. It is hoped that this paper would provide readers with a thorough
reference of this novel, challenging, and thriving field.Comment: Published in Medical Image Analysi
A Survey on Deep Learning in Medical Image Analysis
Deep learning algorithms, in particular convolutional networks, have rapidly
become a methodology of choice for analyzing medical images. This paper reviews
the major deep learning concepts pertinent to medical image analysis and
summarizes over 300 contributions to the field, most of which appeared in the
last year. We survey the use of deep learning for image classification, object
detection, segmentation, registration, and other tasks and provide concise
overviews of studies per application area. Open challenges and directions for
future research are discussed.Comment: Revised survey includes expanded discussion section and reworked
introductory section on common deep architectures. Added missed papers from
before Feb 1st 201
Deep Active Learning for Automatic Mitotic Cell Detection on HEp-2 Specimen Medical Images
Identifying Human Epithelial Type 2 (HEp-2) mitotic cells is a crucial procedure in anti-nuclear antibodies (ANAs) testing, which is the standard protocol for detecting connective tissue diseases (CTD). Due to the low throughput and labor-subjectivity of the ANAs' manual screening test, there is a need to develop a reliable HEp-2 computer-aided diagnosis (CAD) system. The automatic detection of mitotic cells from the microscopic HEp-2 specimen images is an essential step to support the diagnosis process and enhance the throughput of this test. This work proposes a deep active learning (DAL) approach to overcoming the cell labeling challenge. Moreover, deep learning detectors are tailored to automatically identify the mitotic cells directly in the entire microscopic HEp-2 specimen images, avoiding the segmentation step. The proposed framework is validated using the I3A Task-2 dataset over 5-fold cross-validation trials. Using the YOLO predictor, promising mitotic cell prediction results are achieved with an average of 90.011% recall, 88.307% precision, and 81.531% mAP. Whereas, average scores of 86.986% recall, 85.282% precision, and 78.506% mAP are obtained using the Faster R-CNN predictor. Employing the DAL method over four labeling rounds effectively enhances the accuracy of the data annotation, and hence, improves the prediction performance. The proposed framework could be practically applicable to support medical personnel in making rapid and accurate decisions about the mitotic cells' existence
Deep Learning Model Based on ResNet-50 for Beef Quality Classification
Food quality measurement is one of the most essential topics in agriculture and industrial fields. To classify healthy food using computer visual inspection, a new architecture was proposed to classify beef images to specify the rancid and healthy ones. In traditional measurements, the specialists are not able to classify such images, due to the huge number of beef images required to build a deep learning model. In the present study, different images of beef including healthy and rancid cases were collected according to the analysis done by the Laboratory of Food Technology, Faculty of Agriculture, Kafrelsheikh University in January of 2020. The texture analysis of the beef surface of the enrolled images makes it difficult to distinguish between the rancid and healthy images. Moreover, a deep learning approach based on ResNet-50 was presented as a promising classifier to grade and classify the beef images. In this work, a limited number of images were used to present the research problem of image resource limitation; eight healthy images and ten rancid beef images. This number of images is not sufficient to be retrained using deep learning approaches. Thus, Generative Adversarial Network (GAN) was proposed to augment the enrolled images to produce one hundred eighty images. The results obtained based on ResNet-50 classification achieve accuracy of 96.03%, 91.67%, and 88.89% in the training, testing, and validation phases, respectively. Furthermore, a comparison of the current model (ResNet-50) with the classical and deep learning architecture is made to demonstrate the efficiency of ResNet-50, in image classification
On Improving Generalization of CNN-Based Image Classification with Delineation Maps Using the CORF Push-Pull Inhibition Operator
Deployed image classification pipelines are typically dependent on the images captured in real-world environments. This means that images might be affected by different sources of perturbations (e.g. sensor noise in low-light environments). The main challenge arises by the fact that image quality directly impacts the reliability and consistency of classification tasks. This challenge has, hence, attracted wide interest within the computer vision communities. We propose a transformation step that attempts to enhance the generalization ability of CNN models in the presence of unseen noise in the test set. Concretely, the delineation maps of given images are determined using the CORF push-pull inhibition operator. Such an operation transforms an input image into a space that is more robust to noise before being processed by a CNN. We evaluated our approach on the Fashion MNIST data set with an AlexNet model. It turned out that the proposed CORF-augmented pipeline achieved comparable results on noise-free images to those of a conventional AlexNet classification model without CORF delineation maps, but it consistently achieved significantly superior performance on test images perturbed with different levels of Gaussian and uniform noise
Medical Image Analysis using Deep Relational Learning
In the past ten years, with the help of deep learning, especially the rapid
development of deep neural networks, medical image analysis has made remarkable
progress. However, how to effectively use the relational information between
various tissues or organs in medical images is still a very challenging
problem, and it has not been fully studied. In this thesis, we propose two
novel solutions to this problem based on deep relational learning. First, we
propose a context-aware fully convolutional network that effectively models
implicit relation information between features to perform medical image
segmentation. The network achieves the state-of-the-art segmentation results on
the Multi Modal Brain Tumor Segmentation 2017 (BraTS2017) and Multi Modal Brain
Tumor Segmentation 2018 (BraTS2018) data sets. Subsequently, we propose a new
hierarchical homography estimation network to achieve accurate medical image
mosaicing by learning the explicit spatial relationship between adjacent
frames. We use the UCL Fetoscopy Placenta dataset to conduct experiments and
our hierarchical homography estimation network outperforms the other
state-of-the-art mosaicing methods while generating robust and meaningful
mosaicing result on unseen frames.Comment: arXiv admin note: substantial text overlap with arXiv:2007.0778
- …