1,674 research outputs found
Automatic Liver Segmentation Using an Adversarial Image-to-Image Network
Automatic liver segmentation in 3D medical images is essential in many
clinical applications, such as pathological diagnosis of hepatic diseases,
surgical planning, and postoperative assessment. However, it is still a very
challenging task due to the complex background, fuzzy boundary, and various
appearance of liver. In this paper, we propose an automatic and efficient
algorithm to segment liver from 3D CT volumes. A deep image-to-image network
(DI2IN) is first deployed to generate the liver segmentation, employing a
convolutional encoder-decoder architecture combined with multi-level feature
concatenation and deep supervision. Then an adversarial network is utilized
during training process to discriminate the output of DI2IN from ground truth,
which further boosts the performance of DI2IN. The proposed method is trained
on an annotated dataset of 1000 CT volumes with various different scanning
protocols (e.g., contrast and non-contrast, various resolution and position)
and large variations in populations (e.g., ages and pathology). Our approach
outperforms the state-of-the-art solutions in terms of segmentation accuracy
and computing efficiency.Comment: Accepted by MICCAI 201
Task Driven Generative Modeling for Unsupervised Domain Adaptation: Application to X-ray Image Segmentation
Automatic parsing of anatomical objects in X-ray images is critical to many
clinical applications in particular towards image-guided invention and workflow
automation. Existing deep network models require a large amount of labeled
data. However, obtaining accurate pixel-wise labeling in X-ray images relies
heavily on skilled clinicians due to the large overlaps of anatomy and the
complex texture patterns. On the other hand, organs in 3D CT scans preserve
clearer structures as well as sharper boundaries and thus can be easily
delineated. In this paper, we propose a novel model framework for learning
automatic X-ray image parsing from labeled CT scans. Specifically, a Dense
Image-to-Image network (DI2I) for multi-organ segmentation is first trained on
X-ray like Digitally Reconstructed Radiographs (DRRs) rendered from 3D CT
volumes. Then we introduce a Task Driven Generative Adversarial Network
(TD-GAN) architecture to achieve simultaneous style transfer and parsing for
unseen real X-ray images. TD-GAN consists of a modified cycle-GAN substructure
for pixel-to-pixel translation between DRRs and X-ray images and an added
module leveraging the pre-trained DI2I to enforce segmentation consistency. The
TD-GAN framework is general and can be easily adapted to other learning tasks.
In the numerical experiments, we validate the proposed model on 815 DRRs and
153 topograms. While the vanilla DI2I without any adaptation fails completely
on segmenting the topograms, the proposed model does not require any topogram
labels and is able to provide a promising average dice of 85% which achieves
the same level accuracy of supervised training (88%)
- …