1,319 research outputs found
Brain Tumor Synthetic Segmentation in 3D Multimodal MRI Scans
The magnetic resonance (MR) analysis of brain tumors is widely used for
diagnosis and examination of tumor subregions. The overlapping area among the
intensity distribution of healthy, enhancing, non-enhancing, and edema regions
makes the automatic segmentation a challenging task. Here, we show that a
convolutional neural network trained on high-contrast images can transform the
intensity distribution of brain lesions in its internal subregions.
Specifically, a generative adversarial network (GAN) is extended to synthesize
high-contrast images. A comparison of these synthetic images and real images of
brain tumor tissue in MR scans showed significant segmentation improvement and
decreased the number of real channels for segmentation. The synthetic images
are used as a substitute for real channels and can bypass real modalities in
the multimodal brain tumor segmentation framework. Segmentation results on
BraTS 2019 dataset demonstrate that our proposed approach can efficiently
segment the tumor areas. In the end, we predict patient survival time based on
volumetric features of the tumor subregions as well as the age of each case
through several regression models
Convolutional 3D to 2D Patch Conversion for Pixel-wise Glioma Segmentation in MRI Scans
Structural magnetic resonance imaging (MRI) has been widely utilized for
analysis and diagnosis of brain diseases. Automatic segmentation of brain
tumors is a challenging task for computer-aided diagnosis due to low-tissue
contrast in the tumor subregions. To overcome this, we devise a novel
pixel-wise segmentation framework through a convolutional 3D to 2D MR patch
conversion model to predict class labels of the central pixel in the input
sliding patches. Precisely, we first extract 3D patches from each modality to
calibrate slices through the squeeze and excitation (SE) block. Then, the
output of the SE block is fed directly into subsequent bottleneck layers to
reduce the number of channels. Finally, the calibrated 2D slices are
concatenated to obtain multimodal features through a 2D convolutional neural
network (CNN) for prediction of the central pixel. In our architecture, both
local inter-slice and global intra-slice features are jointly exploited to
predict class label of the central voxel in a given patch through the 2D CNN
classifier. We implicitly apply all modalities through trainable parameters to
assign weights to the contributions of each sequence for segmentation.
Experimental results on the segmentation of brain tumors in multimodal MRI
scans (BraTS'19) demonstrate that our proposed method can efficiently segment
the tumor regions
The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS)
In this paper we report the set-up and results of the Multimodal Brain Tumor Image Segmentation Benchmark (BRATS) organized in conjunction with the MICCAI 2012 and 2013 conferences. Twenty state-of-the-art tumor segmentation algorithms were applied to a set of 65 multi-contrast MR scans of low-and high-grade glioma patients-manually annotated by up to four raters-and to 65 comparable scans generated using tumor image simulation software. Quantitative evaluations revealed considerable disagreement between the human raters in segmenting various tumor sub-regions (Dice scores in the range 74%-85%), illustrating the difficulty of this task. We found that different algorithms worked best for different sub-regions (reaching performance comparable to human inter-rater variability), but that no single algorithm ranked in the top for all sub-regions simultaneously. Fusing several good algorithms using a hierarchical majority vote yielded segmentations that consistently ranked above all individual algorithms, indicating remaining opportunities for further methodological improvements. The BRATS image data and manual annotations continue to be publicly available through an online evaluation system as an ongoing benchmarking resource
- …