329 research outputs found
Semi-automatic Liver Tumor Segmentation in Dynamic Contrast-Enhanced CT Scans Using Random Forests and Supervoxels
International audiencePre-operative locoregional treatments (PLT) delay the tumor progression by necrosis for patients with hepato-cellular carcinoma (HCC). Toward an efficient evaluation of PLT response, we address the estimation of liver tumor necrosis (TN) from CT scans. The TN rate could shortly supplant standard criteria (RECIST, mRECIST, EASL or WHO) since it has recently shown higher correlation to survival rates. To overcome the inter-expert variability induced by visual qualitative assessment, we propose a semi-automatic method that requires weak interaction efforts to segment parenchyma, tumoral active and necrotic tissues. By combining SLIC supervoxels and random decision forest, it involves discriminative multi-phase cluster-wise features extracted from registered dynamic contrast-enhanced CT scans. Quantitative assessment on expert groundtruth annotations confirms the benefits of exploiting multi-phase information from semantic regions to accurately segment HCC liver tumors
Patch-based segmentation with spatial context for medical image analysis
Accurate segmentations in medical imaging form a crucial role in many applications from pa-
tient diagnosis to population studies. As the amount of data generated from medical images
increases, the ability to perform this task without human intervention becomes ever more de-
sirable. One approach, known broadly as atlas-based segmentation, is to propagate labels from
images which have already been manually labelled by clinical experts. Methods using this ap-
proach have been shown to be e ective in many applications, demonstrating great potential for
automatic labelling of large datasets. However, these methods usually require the use of image
registration and are dependent on the outcome of the registration. Any registrations errors
that occur are also propagated to the segmentation process and are likely to have an adverse
e ect on segmentation accuracy. Recently, patch-based methods have been shown to allow a
relaxation of the required image alignment, whilst achieving similar results. In general, these
methods label each voxel of a target image by comparing the image patch centred on the voxel
with neighbouring patches from an atlas library and assigning the most likely label according
to the closest matches. The main contributions of this thesis focuses around this approach
in providing accurate segmentation results whilst minimising the dependency on registration
quality. In particular, this thesis proposes a novel kNN patch-based segmentation framework,
which utilises both intensity and spatial information, and explore the use of spatial context in
a diverse range of applications. The proposed methods extend the potential for patch-based
segmentation to tolerate registration errors by rede ning the \locality" for patch selection and
comparison, whilst also allowing similar looking patches from di erent anatomical structures
to be di erentiated. The methods are evaluated on a wide variety of image datasets, ranging
from the brain to the knees, demonstrating its potential with results which are competitive to
state-of-the-art techniques.Open Acces
Computational Anatomy for Multi-Organ Analysis in Medical Imaging: A Review
The medical image analysis field has traditionally been focused on the
development of organ-, and disease-specific methods. Recently, the interest in
the development of more 20 comprehensive computational anatomical models has
grown, leading to the creation of multi-organ models. Multi-organ approaches,
unlike traditional organ-specific strategies, incorporate inter-organ relations
into the model, thus leading to a more accurate representation of the complex
human anatomy. Inter-organ relations are not only spatial, but also functional
and physiological. Over the years, the strategies 25 proposed to efficiently
model multi-organ structures have evolved from the simple global modeling, to
more sophisticated approaches such as sequential, hierarchical, or machine
learning-based models. In this paper, we present a review of the state of the
art on multi-organ analysis and associated computation anatomy methodology. The
manuscript follows a methodology-based classification of the different
techniques 30 available for the analysis of multi-organs and multi-anatomical
structures, from techniques using point distribution models to the most recent
deep learning-based approaches. With more than 300 papers included in this
review, we reflect on the trends and challenges of the field of computational
anatomy, the particularities of each anatomical region, and the potential of
multi-organ analysis to increase the impact of 35 medical imaging applications
on the future of healthcare.Comment: Paper under revie
Contour-Driven Atlas-Based Segmentation
We propose new methods for automatic segmentation of images based on an atlas of manually labeled scans and contours in the image. First, we introduce a Bayesian framework for creating initial label maps from manually annotated training images. Within this framework, we model various registration- and patch-based segmentation techniques by changing the deformation field prior. Second, we perform contour-driven regression on the created label maps to refine the segmentation. Image contours and image parcellations give rise to non-stationary kernel functions that model the relationship between image locations. Setting the kernel to the covariance function in a Gaussian process establishes a distribution over label maps supported by image structures. Maximum a posteriori estimation of the distribution over label maps conditioned on the outcome of the atlas-based segmentation yields the refined segmentation. We evaluate the segmentation in two clinical applications: the segmentation of parotid glands in head and neck CT scans and the segmentation of the left atrium in cardiac MR angiography images
Stacked fully convolutional networks with multi-channel learning: application to medical image segmentation
The automated segmentation of regions of interest (ROIs) in medical imaging is the fundamental requirement for the derivation of high-level semantics for image analysis in clinical decision support systems. Traditional segmentation approaches such as region-based depend heavily upon hand-crafted features and a priori knowledge of the user. As such, these methods are difficult to adopt within a clinical environment. Recently, methods based on fully convolutional networks (FCN) have achieved great success in the segmentation of general images. FCNs leverage a large labeled dataset to hierarchically learn the features that best correspond to the shallow appearance as well as the deep semantics of the images. However, when applied to medical images, FCNs usually produce coarse ROI detection and poor boundary definitions primarily due to the limited number of labeled training data and limited constraints of label agreement among neighboring similar pixels. In this paper, we propose a new stacked FCN architecture with multi-channel learning (SFCN-ML). We embed the FCN in a stacked architecture to learn the foreground ROI features and background non-ROI features separately and then integrate these different channels to produce the final segmentation result. In contrast to traditional FCN methods, our SFCN-ML architecture enables the visual attributes and semantics derived from both the fore- and background channels to be iteratively learned and inferred. We conducted extensive experiments on three public datasets with a variety of visual challenges. Our results show that our SFCN-ML is more effective and robust than a routine FCN and its variants, and other state-of-the-art methods
A Survey on Deep Learning in Medical Image Analysis
Deep learning algorithms, in particular convolutional networks, have rapidly
become a methodology of choice for analyzing medical images. This paper reviews
the major deep learning concepts pertinent to medical image analysis and
summarizes over 300 contributions to the field, most of which appeared in the
last year. We survey the use of deep learning for image classification, object
detection, segmentation, registration, and other tasks and provide concise
overviews of studies per application area. Open challenges and directions for
future research are discussed.Comment: Revised survey includes expanded discussion section and reworked
introductory section on common deep architectures. Added missed papers from
before Feb 1st 201
Nephroblastoma analysis in MRI images
The annotation of the tumour from medical scans is a crucial step in nephroblastoma treatment. Therefore, an accurate and reliable segmentation method is needed to facilitate the evaluation and the treatments of the tumour. The proposed method serves this purpose by performing the segmentation of nephroblastoma in MRI scans. The segmentation is performed by adapting and a 2D free hand drawing tool to select a region of interest in the scan slices. Results from 24 patients show a mean root-mean-square error of 0.0481±0.0309, an average Dice coefficient of 0.9060±0.0549 and an average accuracy of 99.59% ±0.0039. Thus the proposed method demonstrated an effective agreement with manual annotations
- …