704 research outputs found
2.5D multi-view gait recognition based on point cloud registration
This paper presents a method for modeling a 2.5-dimensional (2.5D) human body and extracting the gait features for identifying the human subject. To achieve view-invariant gait recognition, a multi-view synthesizing method based on point cloud registration (MVSM) to generate multi-view training galleries is proposed. The concept of a density and curvature-based Color Gait Curvature Image is introduced to map 2.5D data onto a 2D space to enable data dimension reduction by discrete cosine transform and 2D principle component analysis. Gait recognition is achieved via a 2.5D view-invariant gait recognition method based on point cloud registration. Experimental results on the in-house database captured by a Microsoft Kinect camera show a significant performance gain when using MVSM
Explainable Anatomical Shape Analysis through Deep Hierarchical Generative Models
Quantification of anatomical shape changes currently relies on scalar global indexes which are largely insensitive to regional or asymmetric modifications. Accurate assessment of pathology-driven anatomical remodeling is a crucial step for the diagnosis and treatment of many conditions. Deep learning approaches have recently achieved wide success in the analysis of medical images, but they lack interpretability in the feature extraction and decision processes. In this work, we propose a new interpretable deep learning model for shape analysis. In particular, we exploit deep generative networks to model a population of anatomical segmentations through a hierarchy of conditional latent variables. At the highest level of this hierarchy, a two-dimensional latent space is simultaneously optimised to discriminate distinct clinical conditions, enabling the direct visualisation of the classification space. Moreover, the anatomical variability encoded by this discriminative latent space can be visualised in the segmentation space thanks to the generative properties of the model, making the classification task transparent. This approach yielded high accuracy in the categorisation of healthy and remodelled left ventricles when tested on unseen segmentations from our own multi-centre dataset as well as in an external validation set, and on hippocampi from healthy controls and patients with Alzheimer's disease when tested on ADNI data. More importantly, it enabled the visualisation in three-dimensions of both global and regional anatomical features which better discriminate between the conditions under exam. The proposed approach scales effectively to large populations, facilitating high-throughput analysis of normal anatomy and pathology in large-scale studies of volumetric imaging
GridNet with automatic shape prior registration for automatic MRI cardiac segmentation
In this paper, we propose a fully automatic MRI cardiac segmentation method
based on a novel deep convolutional neural network (CNN) designed for the 2017
ACDC MICCAI challenge. The novelty of our network comes with its embedded shape
prior and its loss function tailored to the cardiac anatomy. Our model includes
a cardiac centerof-mass regression module which allows for an automatic shape
prior registration. Also, since our method processes raw MR images without any
manual preprocessing and/or image cropping, our CNN learns both high-level
features (useful to distinguish the heart from other organs with a similar
shape) and low-level features (useful to get accurate segmentation results).
Those features are learned with a multi-resolution conv-deconv "grid"
architecture which can be seen as an extension of the U-Net. Experimental
results reveal that our method can segment the left and right ventricles as
well as the myocardium from a 3D MRI cardiac volume in 0.4 second with an
average Dice coefficient of 0.90 and an average Hausdorff distance of 10.4 mm.Comment: 8 pages, 1 tables, 2 figure
Computational Methods for Segmentation of Multi-Modal Multi-Dimensional Cardiac Images
Segmentation of the heart structures helps compute the cardiac contractile function quantified via the systolic and diastolic volumes, ejection fraction, and myocardial mass, representing a reliable diagnostic value. Similarly, quantification of the myocardial mechanics throughout the cardiac cycle, analysis of the activation patterns in the heart via electrocardiography (ECG) signals, serve as good cardiac diagnosis indicators. Furthermore, high quality anatomical models of the heart can be used in planning and guidance of minimally invasive interventions under the assistance of image guidance.
The most crucial step for the above mentioned applications is to segment the ventricles and myocardium from the acquired cardiac image data. Although the manual delineation of the heart structures is deemed as the gold-standard approach, it requires significant time and effort, and is highly susceptible to inter- and intra-observer variability. These limitations suggest a need for fast, robust, and accurate semi- or fully-automatic segmentation algorithms. However, the complex motion and anatomy of the heart, indistinct borders due to blood flow, the presence of trabeculations, intensity inhomogeneity, and various other imaging artifacts, makes the segmentation task challenging.
In this work, we present and evaluate segmentation algorithms for multi-modal, multi-dimensional cardiac image datasets. Firstly, we segment the left ventricle (LV) blood-pool from a tri-plane 2D+time trans-esophageal (TEE) ultrasound acquisition using local phase based filtering and graph-cut technique, propagate the segmentation throughout the cardiac cycle using non-rigid registration-based motion extraction, and reconstruct the 3D LV geometry. Secondly, we segment the LV blood-pool and myocardium from an open-source 4D cardiac cine Magnetic Resonance Imaging (MRI) dataset by incorporating average atlas based shape constraint into the graph-cut framework and iterative segmentation refinement. The developed fast and robust framework is further extended to perform right ventricle (RV) blood-pool segmentation from a different open-source 4D cardiac cine MRI dataset. Next, we employ convolutional neural network based multi-task learning framework to segment the myocardium and regress its area, simultaneously, and show that segmentation based computation of the myocardial area is significantly better than that regressed directly from the network, while also being more interpretable. Finally, we impose a weak shape constraint via multi-task learning framework in a fully convolutional network and show improved segmentation performance for LV, RV and myocardium across healthy and pathological cases, as well as, in the challenging apical and basal slices in two open-source 4D cardiac cine MRI datasets.
We demonstrate the accuracy and robustness of the proposed segmentation methods by comparing the obtained results against the provided gold-standard manual segmentations, as well as with other competing segmentation methods
Multi-resolution dental image registration based on genetic algorithm
The Automated Dental Identification System (ADIS) is a Post Mortem Dental Identification System. This thesis presents dental image registration, required for the preprocessing steps of the image comparison component of ADIS. We proposed a multi resolution dental image registration based on genetic algorithms. The main objective of this research is to develop techniques for registration of extracted subject regions of interest with corresponding reference regions of interest.;We investigated and implemented registration using two multi resolution techniques namely image sub sampling and wavelet decomposition. Multi resolution techniques help in the reduction of search data since initial registration is carried at lower levels and results are updated as the levels of resolutions increase. We adopted edges as image features that needed to be aligned. Affine transformations were selected to transform the subject dental region of interest to achieve better alignment with the reference region of interest. These transformations are known to capture complex image distortions. The similarity between subject and reference image has been computed using Oriented Hausdorff Similarity measure that is robust to severe noise and image degradations. A genetic algorithm was adopted to search for the best transformation parameters that give maximum similarity score.;Testing results show that the developed registration algorithm yielded reasonable results in accuracy for dental test cases that contained slight misalignments. The relative percentage errors between the known and estimated transformation parameters were less than 20% with a termination criterion of a ten minute time limit. Further research is needed for dental cases that contain high degree of misalignment, noise and distortions
Deep convolutional neural networks for segmenting 3D in vivo multiphoton images of vasculature in Alzheimer disease mouse models
The health and function of tissue rely on its vasculature network to provide
reliable blood perfusion. Volumetric imaging approaches, such as multiphoton
microscopy, are able to generate detailed 3D images of blood vessels that could
contribute to our understanding of the role of vascular structure in normal
physiology and in disease mechanisms. The segmentation of vessels, a core image
analysis problem, is a bottleneck that has prevented the systematic comparison
of 3D vascular architecture across experimental populations. We explored the
use of convolutional neural networks to segment 3D vessels within volumetric in
vivo images acquired by multiphoton microscopy. We evaluated different network
architectures and machine learning techniques in the context of this
segmentation problem. We show that our optimized convolutional neural network
architecture, which we call DeepVess, yielded a segmentation accuracy that was
better than both the current state-of-the-art and a trained human annotator,
while also being orders of magnitude faster. To explore the effects of aging
and Alzheimer's disease on capillaries, we applied DeepVess to 3D images of
cortical blood vessels in young and old mouse models of Alzheimer's disease and
wild type littermates. We found little difference in the distribution of
capillary diameter or tortuosity between these groups, but did note a decrease
in the number of longer capillary segments () in aged animals as
compared to young, in both wild type and Alzheimer's disease mouse models.Comment: 34 pages, 9 figure
PWD-3DNet: A Deep Learning-Based Fully-Automated Segmentation of Multiple Structures on Temporal Bone CT Scans
The temporal bone is a part of the lateral skull surface that contains organs responsible for hearing and balance. Mastering surgery of the temporal bone is challenging because of this complex and microscopic three-dimensional anatomy. Segmentation of intra-temporal anatomy based on computed tomography (CT) images is necessary for applications such as surgical training and rehearsal, amongst others. However, temporal bone segmentation is challenging due to the similar intensities and complicated anatomical relationships among critical structures, undetectable small structures on standard clinical CT, and the amount of time required for manual segmentation. This paper describes a single multi-class deep learning-based pipeline as the first fully automated algorithm for segmenting multiple temporal bone structures from CT volumes, including the sigmoid sinus, facial nerve, inner ear, malleus, incus, stapes, internal carotid artery and internal auditory canal. The proposed fully convolutional network, PWD-3DNet, is a patch-wise densely connected (PWD) three-dimensional (3D) network. The accuracy and speed of the proposed algorithm was shown to surpass current manual and semi-automated segmentation techniques. The experimental results yielded significantly high Dice similarity scores and low Hausdorff distances for all temporal bone structures with an average of 86% and 0.755 millimeter (mm), respectively. We illustrated that overlapping in the inference sub-volumes improves the segmentation performance. Moreover, we proposed augmentation layers by using samples with various transformations and image artefacts to increase the robustness of PWD-3DNet against image acquisition protocols, such as smoothing caused by soft tissue scanner settings and larger voxel sizes used for radiation reduction. The proposed algorithm was tested on low-resolution CTs acquired by another center with different scanner parameters than the ones used to create the algorithm and shows potential for application beyond the particular training data used in the study
Learning a Generative Motion Model from Image Sequences based on a Latent Motion Matrix
We propose to learn a probabilistic motion model from a sequence of images
for spatio-temporal registration. Our model encodes motion in a low-dimensional
probabilistic space - the motion matrix - which enables various motion analysis
tasks such as simulation and interpolation of realistic motion patterns
allowing for faster data acquisition and data augmentation. More precisely, the
motion matrix allows to transport the recovered motion from one subject to
another simulating for example a pathological motion in a healthy subject
without the need for inter-subject registration. The method is based on a
conditional latent variable model that is trained using amortized variational
inference. This unsupervised generative model follows a novel multivariate
Gaussian process prior and is applied within a temporal convolutional network
which leads to a diffeomorphic motion model. Temporal consistency and
generalizability is further improved by applying a temporal dropout training
scheme. Applied to cardiac cine-MRI sequences, we show improved registration
accuracy and spatio-temporally smoother deformations compared to three
state-of-the-art registration algorithms. Besides, we demonstrate the model's
applicability for motion analysis, simulation and super-resolution by an
improved motion reconstruction from sequences with missing frames compared to
linear and cubic interpolation.Comment: accepted at IEEE TM
- …