6,196 research outputs found
3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation
This paper introduces a network for volumetric segmentation that learns from
sparsely annotated volumetric images. We outline two attractive use cases of
this method: (1) In a semi-automated setup, the user annotates some slices in
the volume to be segmented. The network learns from these sparse annotations
and provides a dense 3D segmentation. (2) In a fully-automated setup, we assume
that a representative, sparsely annotated training set exists. Trained on this
data set, the network densely segments new volumetric images. The proposed
network extends the previous u-net architecture from Ronneberger et al. by
replacing all 2D operations with their 3D counterparts. The implementation
performs on-the-fly elastic deformations for efficient data augmentation during
training. It is trained end-to-end from scratch, i.e., no pre-trained network
is required. We test the performance of the proposed method on a complex,
highly variable 3D structure, the Xenopus kidney, and achieve good results for
both use cases.Comment: Conditionally accepted for MICCAI 201
Review on Computer Vision in Gastric Cancer: Potential Efficient Tools for Diagnosis
Rapid diagnosis of gastric cancer is a great challenge for clinical doctors.
Dramatic progress of computer vision on gastric cancer has been made recently
and this review focuses on advances during the past five years. Different
methods for data generation and augmentation are presented, and various
approaches to extract discriminative features compared and evaluated.
Classification and segmentation techniques are carefully discussed for
assisting more precise diagnosis and timely treatment. For classification,
various methods have been developed to better proceed specific images, such as
images with rotation and estimated real-timely (endoscopy), high resolution
images (histopathology), low diagnostic accuracy images (X-ray), poor contrast
images of the soft-tissue with cavity (CT) or those images with insufficient
annotation. For detection and segmentation, traditional methods and machine
learning methods are compared. Application of those methods will greatly reduce
the labor and time consumption for the diagnosis of gastric cancers
Stack-U-Net: Refinement Network for Image Segmentation on the Example of Optic Disc and Cup
In this work, we propose a special cascade network for image segmentation,
which is based on the U-Net networks as building blocks and the idea of the
iterative refinement. The model was mainly applied to achieve higher
recognition quality for the task of finding borders of the optic disc and cup,
which are relevant to the presence of glaucoma. Compared to a single U-Net and
the state-of-the-art methods for the investigated tasks, very high segmentation
quality has been achieved without a need for increasing the volume of datasets.
Our experiments include comparison with the best-known methods on publicly
available databases DRIONS-DB, RIM-ONE v.3, DRISHTI-GS, and evaluation on a
private data set collected in collaboration with University of California San
Francisco Medical School. The analysis of the architecture details is
presented, and it is argued that the model can be employed for a broad scope of
image segmentation problems of similar nature
Deep Learning based Computer-Aided Diagnosis Systems for Diabetic Retinopathy: A Survey
Diabetic retinopathy (DR) results in vision loss if not treated early. A
computer-aided diagnosis (CAD) system based on retinal fundus images is an
efficient and effective method for early DR diagnosis and assisting experts. A
computer-aided diagnosis (CAD) system involves various stages like detection,
segmentation and classification of lesions in fundus images. Many traditional
machine-learning (ML) techniques based on hand-engineered features have been
introduced. The recent emergence of deep learning (DL) and its decisive victory
over traditional ML methods for various applications motivated the researchers
to employ it for DR diagnosis, and many deep-learning-based methods have been
introduced. In this paper, we review these methods, highlighting their pros and
cons. In addition, we point out the challenges to be addressed in designing and
learning about efficient, effective and robust deep-learning algorithms for
various problems in DR diagnosis and draw attention to directions for future
research
Semi-supervised Skin Lesion Segmentation via Transformation Consistent Self-ensembling Model
Automatic skin lesion segmentation on dermoscopic images is an essential
component in computer-aided diagnosis of melanoma. Recently, many fully
supervised deep learning based methods have been proposed for automatic skin
lesion segmentation. However, these approaches require massive pixel-wise
annotation from experienced dermatologists, which is very costly and
time-consuming. In this paper, we present a novel semi-supervised method for
skin lesion segmentation by leveraging both labeled and unlabeled data. The
network is optimized by the weighted combination of a common supervised loss
for labeled inputs only and a regularization loss for both labeled and
unlabeled data. In this paper, we present a novel semi-supervised method for
skin lesion segmentation, where the network is optimized by the weighted
combination of a common supervised loss for labeled inputs only and a
regularization loss for both labeled and unlabeled data. Our method encourages
a consistent prediction for unlabeled images using the outputs of the
network-in-training under different regularizations, so that it can utilize the
unlabeled data. To utilize the unlabeled data, our method encourages the
consistent predictions of the network-in-training for the same input under
different regularizations. Aiming for the semi-supervised segmentation problem,
we enhance the effect of regularization for pixel-level predictions by
introducing a transformation, including rotation and flipping, consistent
scheme in our self-ensembling model. With only 300 labeled training samples,
our method sets a new record on the benchmark of the International Skin Imaging
Collaboration (ISIC) 2017 skin lesion segmentation challenge. Such a result
clearly surpasses fully-supervised state-of-the-arts that are trained with 2000
labeled data.Comment: BMVC 201
Transformation Consistent Self-ensembling Model for Semi-supervised Medical Image Segmentation
Deep convolutional neural networks have achieved remarkable progress on a
variety of medical image computing tasks. A common problem when applying
supervised deep learning methods to medical images is the lack of labeled data,
which is very expensive and time-consuming to be collected. In this paper, we
present a novel semi-supervised method for medical image segmentation, where
the network is optimized by the weighted combination of a common supervised
loss for labeled inputs only and a regularization loss for both labeled and
unlabeled data. To utilize the unlabeled data, our method encourages the
consistent predictions of the network-in-training for the same input under
different regularizations. Aiming for the semi-supervised segmentation problem,
we enhance the effect of regularization for pixel-level predictions by
introducing a transformation, including rotation and flipping, consistent
scheme in our self-ensembling model. With the aim of semi-supervised
segmentation tasks, we introduce a transformation consistent strategy in our
self-ensembling model to enhance the regularization effect for pixel-level
predictions. We have extensively validated the proposed semi-supervised method
on three typical yet challenging medical image segmentation tasks: (i) skin
lesion segmentation from dermoscopy images on International Skin Imaging
Collaboration (ISIC) 2017 dataset, (ii) optic disc segmentation from fundus
images on Retinal Fundus Glaucoma Challenge (REFUGE) dataset, and (iii) liver
segmentation from volumetric CT scans on Liver Tumor Segmentation Challenge
(LiTS) dataset. Compared to the state-of-the-arts, our proposed method shows
superior segmentation performance on challenging 2D/3D medical images,
demonstrating the effectiveness of our semi-supervised method for medical image
segmentation.Comment: Accept at IEEE Transactions on Neural Networks and Learning System
Deep Learning Convolutional Networks for Multiphoton Microscopy Vasculature Segmentation
Recently there has been an increasing trend to use deep learning frameworks
for both 2D consumer images and for 3D medical images. However, there has been
little effort to use deep frameworks for volumetric vascular segmentation. We
wanted to address this by providing a freely available dataset of 12 annotated
two-photon vasculature microscopy stacks. We demonstrated the use of deep
learning framework consisting both 2D and 3D convolutional filters (ConvNet).
Our hybrid 2D-3D architecture produced promising segmentation result. We
derived the architectures from Lee et al. who used the ZNN framework initially
designed for electron microscope image segmentation. We hope that by sharing
our volumetric vasculature datasets, we will inspire other researchers to
experiment with vasculature dataset and improve the used network architectures.Comment: 23 pages, 10 figure
Data Efficient and Weakly Supervised Computational Pathology on Whole Slide Images
The rapidly emerging field of computational pathology has the potential to
enable objective diagnosis, therapeutic response prediction and identification
of new morphological features of clinical relevance. However, deep
learning-based computational pathology approaches either require manual
annotation of gigapixel whole slide images (WSIs) in fully-supervised settings
or thousands of WSIs with slide-level labels in a weakly-supervised setting.
Moreover, whole slide level computational pathology methods also suffer from
domain adaptation and interpretability issues. These challenges have prevented
the broad adaptation of computational pathology for clinical and research
purposes. Here we present CLAM - Clustering-constrained attention multiple
instance learning, an easy-to-use, high-throughput, and interpretable WSI-level
processing and learning method that only requires slide-level labels while
being data efficient, adaptable and capable of handling multi-class subtyping
problems. CLAM is a deep-learning-based weakly-supervised method that uses
attention-based learning to automatically identify sub-regions of high
diagnostic value in order to accurately classify the whole slide, while also
utilizing instance-level clustering over the representative regions identified
to constrain and refine the feature space. In three separate analyses, we
demonstrate the data efficiency and adaptability of CLAM and its superior
performance over standard weakly-supervised classification. We demonstrate that
CLAM models are interpretable and can be used to identify well-known and new
morphological features. We further show that models trained using CLAM are
adaptable to independent test cohorts, cell phone microscopy images, and
biopsies. CLAM is a general-purpose and adaptable method that can be used for a
variety of different computational pathology tasks in both clinical and
research settings
Deep Adversarial Training for Multi-Organ Nuclei Segmentation in Histopathology Images
Nuclei segmentation is a fundamental task that is critical for various
computational pathology applications including nuclei morphology analysis, cell
type classification, and cancer grading. Conventional vision-based methods for
nuclei segmentation struggle in challenging cases and deep learning approaches
have proven to be more robust and generalizable. However, CNNs require large
amounts of labeled histopathology data. Moreover, conventional CNN-based
approaches lack structured prediction capabilities which are required to
distinguish overlapping and clumped nuclei. Here, we present an approach to
nuclei segmentation that overcomes these challenges by utilizing a conditional
generative adversarial network (cGAN) trained with synthetic and real data. We
generate a large dataset of H&E training images with perfect nuclei
segmentation labels using an unpaired GAN framework. This synthetic data along
with real histopathology data from six different organs are used to train a
conditional GAN with spectral normalization and gradient penalty for nuclei
segmentation. This adversarial regression framework enforces higher order
consistency when compared to conventional CNN models. We demonstrate that this
nuclei segmentation approach generalizes across different organs, sites,
patients and disease states, and outperforms conventional approaches,
especially in isolating individual and overlapping nuclei
Exploiting Epistemic Uncertainty of Anatomy Segmentation for Anomaly Detection in Retinal OCT
Diagnosis and treatment guidance are aided by detecting relevant biomarkers
in medical images. Although supervised deep learning can perform accurate
segmentation of pathological areas, it is limited by requiring a-priori
definitions of these regions, large-scale annotations, and a representative
patient cohort in the training set. In contrast, anomaly detection is not
limited to specific definitions of pathologies and allows for training on
healthy samples without annotation. Anomalous regions can then serve as
candidates for biomarker discovery. Knowledge about normal anatomical structure
brings implicit information for detecting anomalies. We propose to take
advantage of this property using bayesian deep learning, based on the
assumption that epistemic uncertainties will correlate with anatomical
deviations from a normal training set. A Bayesian U-Net is trained on a
well-defined healthy environment using weak labels of healthy anatomy produced
by existing methods. At test time, we capture epistemic uncertainty estimates
of our model using Monte Carlo dropout. A novel post-processing technique is
then applied to exploit these estimates and transfer their layered appearance
to smooth blob-shaped segmentations of the anomalies. We experimentally
validated this approach in retinal optical coherence tomography (OCT) images,
using weak labels of retinal layers. Our method achieved a Dice index of 0.789
in an independent anomaly test set of age-related macular degeneration (AMD)
cases. The resulting segmentations allowed very high accuracy for separating
healthy and diseased cases with late wet AMD, dry geographic atrophy (GA),
diabetic macular edema (DME) and retinal vein occlusion (RVO). Finally, we
qualitatively observed that our approach can also detect other deviations in
normal scans such as cut edge artifacts.Comment: Accepted for publication in IEEE Transactions on Medical Imaging,
201
- …