4,297 research outputs found
HookNet: multi-resolution convolutional neural networks for semantic segmentation in histopathology whole-slide images
We propose HookNet, a semantic segmentation model for histopathology
whole-slide images, which combines context and details via multiple branches of
encoder-decoder convolutional neural networks. Concentricpatches at multiple
resolutions with different fields of view are used to feed different branches
of HookNet, and intermediate representations are combined via a hooking
mechanism. We describe a framework to design and train HookNet for achieving
high-resolution semantic segmentation and introduce constraints to guarantee
pixel-wise alignment in feature maps during hooking. We show the advantages of
using HookNet in two histopathology image segmentation tasks where tissue type
prediction accuracy strongly depends on contextual information, namely (1)
multi-class tissue segmentation in breast cancer and, (2) segmentation of
tertiary lymphoid structures and germinal centers in lung cancer. Weshow the
superiority of HookNet when compared with single-resolution U-Net models
working at different resolutions as well as with a recently published
multi-resolution model for histopathology image segmentatio
Learning to Segment Breast Biopsy Whole Slide Images
We trained and applied an encoder-decoder model to semantically segment
breast biopsy images into biologically meaningful tissue labels. Since
conventional encoder-decoder networks cannot be applied directly on large
biopsy images and the different sized structures in biopsies present novel
challenges, we propose four modifications: (1) an input-aware encoding block to
compensate for information loss, (2) a new dense connection pattern between
encoder and decoder, (3) dense and sparse decoders to combine multi-level
features, (4) a multi-resolution network that fuses the results of
encoder-decoders run on different resolutions. Our model outperforms a
feature-based approach and conventional encoder-decoders from the literature.
We use semantic segmentations produced with our model in an automated diagnosis
task and obtain higher accuracies than a baseline approach that employs an SVM
for feature-based segmentation, both using the same segmentation-based
diagnostic features.Comment: Added more WSI images in appendi
Iterative annotation to ease neural network training: Specialized machine learning in medical image analysis
Neural networks promise to bring robust, quantitative analysis to medical
fields, but adoption is limited by the technicalities of training these
networks. To address this translation gap between medical researchers and
neural networks in the field of pathology, we have created an intuitive
interface which utilizes the commonly used whole slide image (WSI) viewer,
Aperio ImageScope (Leica Biosystems Imaging, Inc.), for the annotation and
display of neural network predictions on WSIs. Leveraging this, we propose the
use of a human-in-the-loop strategy to reduce the burden of WSI annotation. We
track network performance improvements as a function of iteration and quantify
the use of this pipeline for the segmentation of renal histologic findings on
WSIs. More specifically, we present network performance when applied to
segmentation of renal micro compartments, and demonstrate multi-class
segmentation in human and mouse renal tissue slides. Finally, to show the
adaptability of this technique to other medical imaging fields, we demonstrate
its ability to iteratively segment human prostate glands from radiology imaging
data.Comment: 15 pages, 7 figures, 2 supplemental figures (on the last page
GAN-based Virtual Re-Staining: A Promising Solution for Whole Slide Image Analysis
Histopathological cancer diagnosis is based on visual examination of stained
tissue slides. Hematoxylin and eosin (H\&E) is a standard stain routinely
employed worldwide. It is easy to acquire and cost effective, but cells and
tissue components show low-contrast with varying tones of dark blue and pink,
which makes difficult visual assessments, digital image analysis, and
quantifications. These limitations can be overcome by IHC staining of target
proteins of the tissue slide. IHC provides a selective, high-contrast imaging
of cells and tissue components, but their use is largely limited by a
significantly more complex laboratory processing and high cost. We proposed a
conditional CycleGAN (cCGAN) network to transform the H\&E stained images into
IHC stained images, facilitating virtual IHC staining on the same slide. This
data-driven method requires only a limited amount of labelled data but will
generate pixel level segmentation results. The proposed cCGAN model improves
the original network \cite{zhu_unpaired_2017} by adding category conditions and
introducing two structural loss functions, which realize a multi-subdomain
translation and improve the translation accuracy as well. % need to give
reasons here. Experiments demonstrate that the proposed model outperforms the
original method in unpaired image translation with multi-subdomains. We also
explore the potential of unpaired images to image translation method applied on
other histology images related tasks with different staining techniques
Patch-based Convolutional Neural Network for Whole Slide Tissue Image Classification
Convolutional Neural Networks (CNN) are state-of-the-art models for many
image classification tasks. However, to recognize cancer subtypes
automatically, training a CNN on gigapixel resolution Whole Slide Tissue Images
(WSI) is currently computationally impossible. The differentiation of cancer
subtypes is based on cellular-level visual features observed on image patch
scale. Therefore, we argue that in this situation, training a patch-level
classifier on image patches will perform better than or similar to an
image-level classifier. The challenge becomes how to intelligently combine
patch-level classification results and model the fact that not all patches will
be discriminative. We propose to train a decision fusion model to aggregate
patch-level predictions given by patch-level CNNs, which to the best of our
knowledge has not been shown before. Furthermore, we formulate a novel
Expectation-Maximization (EM) based method that automatically locates
discriminative patches robustly by utilizing the spatial relationships of
patches. We apply our method to the classification of glioma and non-small-cell
lung carcinoma cases into subtypes. The classification accuracy of our method
is similar to the inter-observer agreement between pathologists. Although it is
impossible to train CNNs on WSIs, we experimentally demonstrate using a
comparable non-cancer dataset of smaller images that a patch-based CNN can
outperform an image-based CNN
- …