1,057 research outputs found
A Fixed-Point Model for Pancreas Segmentation in Abdominal CT Scans
Deep neural networks have been widely adopted for automatic organ
segmentation from abdominal CT scans. However, the segmentation accuracy of
some small organs (e.g., the pancreas) is sometimes below satisfaction,
arguably because deep networks are easily disrupted by the complex and variable
background regions which occupies a large fraction of the input volume. In this
paper, we formulate this problem into a fixed-point model which uses a
predicted segmentation mask to shrink the input region. This is motivated by
the fact that a smaller input region often leads to more accurate segmentation.
In the training process, we use the ground-truth annotation to generate
accurate input regions and optimize network weights. On the testing stage, we
fix the network parameters and update the segmentation results in an iterative
manner. We evaluate our approach on the NIH pancreas segmentation dataset, and
outperform the state-of-the-art by more than 4%, measured by the average
Dice-S{\o}rensen Coefficient (DSC). In addition, we report 62.43% DSC in the
worst case, which guarantees the reliability of our approach in clinical
applications.Comment: Accepted to MICCAI 2017 (8 pages, 3 figures
Recurrent Saliency Transformation Network: Incorporating Multi-Stage Visual Cues for Small Organ Segmentation
We aim at segmenting small organs (e.g., the pancreas) from abdominal CT
scans. As the target often occupies a relatively small region in the input
image, deep neural networks can be easily confused by the complex and variable
background. To alleviate this, researchers proposed a coarse-to-fine approach,
which used prediction from the first (coarse) stage to indicate a smaller input
region for the second (fine) stage. Despite its effectiveness, this algorithm
dealt with two stages individually, which lacked optimizing a global energy
function, and limited its ability to incorporate multi-stage visual cues.
Missing contextual information led to unsatisfying convergence in iterations,
and that the fine stage sometimes produced even lower segmentation accuracy
than the coarse stage.
This paper presents a Recurrent Saliency Transformation Network. The key
innovation is a saliency transformation module, which repeatedly converts the
segmentation probability map from the previous iteration as spatial weights and
applies these weights to the current iteration. This brings us two-fold
benefits. In training, it allows joint optimization over the deep networks
dealing with different input scales. In testing, it propagates multi-stage
visual information throughout iterations to improve segmentation accuracy.
Experiments in the NIH pancreas segmentation dataset demonstrate the
state-of-the-art accuracy, which outperforms the previous best by an average of
over 2%. Much higher accuracies are also reported on several small organs in a
larger dataset collected by ourselves. In addition, our approach enjoys better
convergence properties, making it more efficient and reliable in practice.Comment: Accepted to CVPR 2018 (10 pages, 6 figures
A 3D Coarse-to-Fine Framework for Volumetric Medical Image Segmentation
In this paper, we adopt 3D Convolutional Neural Networks to segment
volumetric medical images. Although deep neural networks have been proven to be
very effective on many 2D vision tasks, it is still challenging to apply them
to 3D tasks due to the limited amount of annotated 3D data and limited
computational resources. We propose a novel 3D-based coarse-to-fine framework
to effectively and efficiently tackle these challenges. The proposed 3D-based
framework outperforms the 2D counterpart to a large margin since it can
leverage the rich spatial infor- mation along all three axes. We conduct
experiments on two datasets which include healthy and pathological pancreases
respectively, and achieve the current state-of-the-art in terms of
Dice-S{\o}rensen Coefficient (DSC). On the NIH pancreas segmentation dataset,
we outperform the previous best by an average of over 2%, and the worst case is
improved by 7% to reach almost 70%, which indicates the reliability of our
framework in clinical applications.Comment: 9 pages, 4 figures, Accepted to 3D
DeepOrgan: Multi-level Deep Convolutional Networks for Automated Pancreas Segmentation
Automatic organ segmentation is an important yet challenging problem for
medical image analysis. The pancreas is an abdominal organ with very high
anatomical variability. This inhibits previous segmentation methods from
achieving high accuracies, especially compared to other organs such as the
liver, heart or kidneys. In this paper, we present a probabilistic bottom-up
approach for pancreas segmentation in abdominal computed tomography (CT) scans,
using multi-level deep convolutional networks (ConvNets). We propose and
evaluate several variations of deep ConvNets in the context of hierarchical,
coarse-to-fine classification on image patches and regions, i.e. superpixels.
We first present a dense labeling of local image patches via
and nearest neighbor fusion. Then we describe a regional
ConvNet () that samples a set of bounding boxes around
each image superpixel at different scales of contexts in a "zoom-out" fashion.
Our ConvNets learn to assign class probabilities for each superpixel region of
being pancreas. Last, we study a stacked leveraging
the joint space of CT intensities and the dense
probability maps. Both 3D Gaussian smoothing and 2D conditional random fields
are exploited as structured predictions for post-processing. We evaluate on CT
images of 82 patients in 4-fold cross-validation. We achieve a Dice Similarity
Coefficient of 83.66.3% in training and 71.810.7% in testing.Comment: To be presented at MICCAI 2015 - 18th International Conference on
Medical Computing and Computer Assisted Interventions, Munich, German
An Alarm System For Segmentation Algorithm Based On Shape Model
It is usually hard for a learning system to predict correctly on rare events
that never occur in the training data, and there is no exception for
segmentation algorithms. Meanwhile, manual inspection of each case to locate
the failures becomes infeasible due to the trend of large data scale and
limited human resource. Therefore, we build an alarm system that will set off
alerts when the segmentation result is possibly unsatisfactory, assuming no
corresponding ground truth mask is provided. One plausible solution is to
project the segmentation results into a low dimensional feature space; then
learn classifiers/regressors to predict their qualities. Motivated by this, in
this paper, we learn a feature space using the shape information which is a
strong prior shared among different datasets and robust to the appearance
variation of input data.The shape feature is captured using a Variational
Auto-Encoder (VAE) network that trained with only the ground truth masks.
During testing, the segmentation results with bad shapes shall not fit the
shape prior well, resulting in large loss values. Thus, the VAE is able to
evaluate the quality of segmentation result on unseen data, without using
ground truth. Finally, we learn a regressor in the one-dimensional feature
space to predict the qualities of segmentation results. Our alarm system is
evaluated on several recent state-of-art segmentation algorithms for 3D medical
segmentation tasks. Compared with other standard quality assessment methods,
our system consistently provides more reliable prediction on the qualities of
segmentation results.Comment: Accepted to ICCV 2019 (10 pages, 4 figures
Attention Gated Networks: Learning to Leverage Salient Regions in Medical Images
We propose a novel attention gate (AG) model for medical image analysis that
automatically learns to focus on target structures of varying shapes and sizes.
Models trained with AGs implicitly learn to suppress irrelevant regions in an
input image while highlighting salient features useful for a specific task.
This enables us to eliminate the necessity of using explicit external
tissue/organ localisation modules when using convolutional neural networks
(CNNs). AGs can be easily integrated into standard CNN models such as VGG or
U-Net architectures with minimal computational overhead while increasing the
model sensitivity and prediction accuracy. The proposed AG models are evaluated
on a variety of tasks, including medical image classification and segmentation.
For classification, we demonstrate the use case of AGs in scan plane detection
for fetal ultrasound screening. We show that the proposed attention mechanism
can provide efficient object localisation while improving the overall
prediction performance by reducing false positives. For segmentation, the
proposed architecture is evaluated on two large 3D CT abdominal datasets with
manual annotations for multiple organs. Experimental results show that AG
models consistently improve the prediction performance of the base
architectures across different datasets and training sizes while preserving
computational efficiency. Moreover, AGs guide the model activations to be
focused around salient regions, which provides better insights into how model
predictions are made. The source code for the proposed AG models is publicly
available.Comment: Accepted for Medical Image Analysis (Special Issue on Medical Imaging
with Deep Learning). arXiv admin note: substantial text overlap with
arXiv:1804.03999, arXiv:1804.0533
- …