11,511 research outputs found
Recommended from our members
Deep learning assisted MRI guided attenuation correction in PET
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University LondonPositron emission tomography (PET) is a unique imaging modality that provides physiological
and functional details of the tissue at the molecular level. However, the acquired PET images
have some limitations such as the attenuation. PET attenuation correction is an essential step to
obtain the full potential of PET quantification. With the wide use of hybrid PET/MR scanners,
magnetic resonance (MR) images are used to address the problem of PET attenuation correction.
The MR images segmentation is one simple and robust approach to create pseudo computed
tomography (CT) images, which are used to generate attenuation coefficient maps to correct the
PET attenuation. Recently, deep learning has been proposed and used as a promising technique
to efficiently perform MR and various medical images segmentation.
In this research work, deep learning guided segmentation approaches have been proposed
to enhance the bone class segmentation of MR brain images in order to generate accurate
pseudo-CT images. The first approach has introduced the combination of handcrafted features
with deep learning features to enrich the set of features. Multiresolution analysis techniques,
which generate multiscale and multidirectional coefficients of an image such as contourlet and
shearlet transforms, are applied and combined with deep convolutional neural network (CNN)
features. Different experiments have been conducted to investigate the number of selected
coefficients and the insertion location of the handcrafted features.
The second approach aims at reducing the segmentation algorithm’s complexity while
maintaining the segmentation performance. An attention based convolutional encode-decoder
network has been proposed to adaptively recalibrate the deep network features. This attention based
network consists of two different squeeze and excitation blocks that excite the features
spatially and channel wise. The two blocks are combined sequentially to decrease the number
of network’s parameters and reduces the model complexity. The third approach has been focuses on the application of transfer learning from different MR sequences such as T1 weighted (T1-w) and T2 weighted (T2-w) images. A
pretrained model with T1-w MR sequences is fine tuned to perform the segmentation of T2-w
images. Multiple fine tuning approaches and experiments have been conducted to study the best
fine tuning mechanism that is able to build an efficient segmentation model for both T1-w and
T2-w segmentation. Clinical datasets of fifty patients with different conditions and diagnosis have been
used to carry an objective evaluation to measure the segmentation performance of the results
obtained by the three proposed methods. The first and second approaches have been validated
with other studies in the literature that applied deep network based segmentation technique to
perform MR based attenuation correction for PET images. The proposed methods have shown
an enhancement in the bone segmentation with an increase of dice similarity coefficient (DSC)
from 0.6179 to 0.6567 using an ensemble of CNNs with an improvement percentage of 6.3%.
The proposed excitation-based CNN has decreased the model complexity by decreasing the
number of trainable parameters by more than 46% where less computing resources are required
to train the model. The proposed hybrid transfer learning method has shown its superiority to
build a multi-sequences (T1-w and T2-w) segmentation approach compared to other applied
transfer learning methods especially with the bone class where the DSC is increased from 0.3841
to 0.5393. Moreover, the hybrid transfer learning approach requires less computing time than
transfer learning using open and conservative fine tuning
Combined MR brain segmentation
The paper presents a new approach to segmentation of brain from the MR studies. The method is fully automated, very efficient, and quick. The main point of this algorithm is subtraction of T1 series form T2 series (therefore we called it combined), followed by a few image processing steps.The method has been tested using the data sets from three sources. The results were compared numerically to those produced by experts. They indicate great effectiveness of the presented algorithm
LINKS: Learning-based multi-source IntegratioN frameworK for Segmentation of infant brain images
Segmentation of infant brain MR images is challenging due to insufficient image quality, severe partial volume effect, and ongoing maturation and myelination processes. In the first year of life, the image contrast between white and gray matters of the infant brain undergoes dramatic changes. In particular, the image contrast is inverted around 6-8 months of age, and the white and gray matter tissues are isointense in both T1- and T2-weighted MR images and thus exhibit the extremely low tissue contrast, which poses significant challenges for automated segmentation. Most previous studies used multi-atlas label fusion strategy, which has the limitation of equally treating the different available image modalities and is often computationally expensive. To cope with these limitations, in this paper, we propose a novel learning-based multi-source integration framework for segmentation of infant brain images. Specifically, we employ the random forest technique to effectively integrate features from multi-source images together for tissue segmentation. Here, the multi-source images include initially only the multi-modality (T1, T2 and FA) images and later also the iteratively estimated and refined tissue probability maps of gray matter, white matter, and cerebrospinal fluid. Experimental results on 119 infants show that the proposed method achieves better performance than other state-of-the-art automated segmentation methods. Further validation was performed on the MICCAI grand challenge and the proposed method was ranked top among all competing methods. Moreover, to alleviate the possible anatomical errors, our method can also be combined with an anatomically-constrained multi-atlas labeling approach for further improving the segmentation accuracy
Automatic segmentation of MR brain images with a convolutional neural network
Automatic segmentation in MR brain images is important for quantitative
analysis in large-scale studies with images acquired at all ages.
This paper presents a method for the automatic segmentation of MR brain
images into a number of tissue classes using a convolutional neural network. To
ensure that the method obtains accurate segmentation details as well as spatial
consistency, the network uses multiple patch sizes and multiple convolution
kernel sizes to acquire multi-scale information about each voxel. The method is
not dependent on explicit features, but learns to recognise the information
that is important for the classification based on training data. The method
requires a single anatomical MR image only.
The segmentation method is applied to five different data sets: coronal
T2-weighted images of preterm infants acquired at 30 weeks postmenstrual age
(PMA) and 40 weeks PMA, axial T2- weighted images of preterm infants acquired
at 40 weeks PMA, axial T1-weighted images of ageing adults acquired at an
average age of 70 years, and T1-weighted images of young adults acquired at an
average age of 23 years. The method obtained the following average Dice
coefficients over all segmented tissue classes for each data set, respectively:
0.87, 0.82, 0.84, 0.86 and 0.91.
The results demonstrate that the method obtains accurate segmentations in all
five sets, and hence demonstrates its robustness to differences in age and
acquisition protocol
Reliability of two techniques for assessing cerebral iron deposits from structural MRI
Purpose: To test the reliability of two computational methods for segmenting cerebral iron deposits (IDs) in the ageing brain, given that its measurement in MRI is challenging due to the similar effect produced by other minerals, especially calcium, on T2*-weighted sequences.
Materials and Methods: T1-, T2*-weighted and FLAIR MR brain images obtained at 1.5T from 70 subjects in their early 70s who displayed a wide range of brain IDs were analyzed. The first segmentation method used a multispectral approach based on the fusion of two or more structural sequences registered and mapped in the red/green color space followed by Minimum Variance Quantization. The second method employed a combined thresholding, size and shape analysis using T2*-weighted images augmented with visual information from T1-weighted data.
Results: Both segmentation techniques had high intra- and inter-observer agreement (95 % CI = ± 57 voxels in a range from 0 to 1800), which decreased in subjects with significant microbleeds and/or IDs. However, the thresholding method was more observer dependent in identifying microbleeds and IDs boundaries than the multispectral approach.
Conclusion: Both techniques proved to be in agreement and have good intra- and inter-observer reliability. However, they have limitations, specifically with regard to automation and observer independence, so further work is required to develop fully user-independent methods of identifying cerebral IDs
- …