50,121 research outputs found
Label-driven weakly-supervised learning for multimodal deformable image registration
Spatially aligning medical images from different modalities remains a
challenging task, especially for intraoperative applications that require fast
and robust algorithms. We propose a weakly-supervised, label-driven formulation
for learning 3D voxel correspondence from higher-level label correspondence,
thereby bypassing classical intensity-based image similarity measures. During
training, a convolutional neural network is optimised by outputting a dense
displacement field (DDF) that warps a set of available anatomical labels from
the moving image to match their corresponding counterparts in the fixed image.
These label pairs, including solid organs, ducts, vessels, point landmarks and
other ad hoc structures, are only required at training time and can be
spatially aligned by minimising a cross-entropy function of the warped moving
label and the fixed label. During inference, the trained network takes a new
image pair to predict an optimal DDF, resulting in a fully-automatic,
label-free, real-time and deformable registration. For interventional
applications where large global transformation prevails, we also propose a
neural network architecture to jointly optimise the global- and local
displacements. Experiment results are presented based on cross-validating
registrations of 111 pairs of T2-weighted magnetic resonance images and 3D
transrectal ultrasound images from prostate cancer patients with a total of
over 4000 anatomical labels, yielding a median target registration error of 4.2
mm on landmark centroids and a median Dice of 0.88 on prostate glands.Comment: Accepted to ISBI 201
Accurate Single Image Multi-Modal Camera Pose Estimation
Abstract. A well known problem in photogrammetry and computer vision is the precise and robust determination of camera poses with respect to a given 3D model. In this work we propose a novel multi-modal method for single image camera pose estimation with respect to 3D models with intensity information (e.g., LiDAR data with reflectance information). We utilize a direct point based rendering approach to generate synthetic 2D views from 3D datasets in order to bridge the dimensionality gap. The proposed method then establishes 2D/2D point and local region correspondences based on a novel self-similarity distance measure. Correct correspondences are robustly identified by searching for small regions with a similar geometric relationship of local self-similarities using a Generalized Hough Transform. After backprojection of the generated features into 3D a standard Perspective-n-Points problem is solved to yield an initial camera pose. The pose is then accurately refined using an intensity based 2D/3D registration approach. An evaluation on Vis/IR 2D and airborne and terrestrial 3D datasets shows that the proposed method is applicable to a wide range of different sensor types. In addition, the approach outperforms standard global multi-modal 2D/3D registration approaches based on Mutual Information with respect to robustness and speed. Potential applications are widespread and include for instance multispectral texturing of 3D models, SLAM applications, sensor data fusion and multi-spectral camera calibration and super-resolution applications
Non-rigid registration of 2-D/3-D dynamic data with feature alignment
In this work, we are computing the matching between 2D manifolds and 3D manifolds with temporal constraints, that is we are computing the matching among a time sequence of 2D/3D manifolds. It is solved by mapping all the manifolds to a common domain, then build their matching by composing the forward mapping and the inverse mapping. At first, we solve the matching problem between 2D manifolds with temporal constraints by using mesh-based registration method. We propose a surface parameterization method to compute the mapping between the 2D manifold and the common 2D planar domain. We can compute the matching among the time sequence of deforming geometry data through this common domain. Compared with previous work, our method is independent of the quality of mesh elements and more efficient for the time sequence data. Then we develop a global intensity-based registration method to solve the matching problem between 3D manifolds with temporal constraints. Our method is based on a 4D(3D+T) free-from B-spline deformation model which has both spatial and temporal smoothness. Compared with previous 4D image registration techniques, our method avoids some local minimum. Thus it can be solved faster and achieve better accuracy of landmark point predication. We demonstrate the efficiency of these works on the real applications. The first one is applied to the dynamic face registering and texture mapping. The second one is applied to lung tumor motion tracking in the medical image analysis. In our future work, we are developing more efficient mesh-based 4D registration method. It can be applied to tumor motion estimation and tracking, which can be used to calculate the read dose delivered to the lung and surrounding tissues. Thus this can support the online treatment of lung cancer radiotherapy
Recommended from our members
Endoluminal surface registration for CT colonography using Haustral Fold Matching
Computed Tomographic (CT) colonography is a technique used for the detection of bowel cancer or potentially precancerous polyps. The procedure is performed routinely with the patient both prone and supine to differentiate fixed colonic pathology from mobile faecal residue. Matching corresponding locations is difficult and time consuming for radiologists due to colonic deformations that occur during patient repositioning.
We propose a novel method to establish correspondence between the two acquisitions automatically. The problem is first simplified by detecting haustral folds using a graph cut method applied to a curvature-based metric applied to a surface mesh generated from segmentation of the colonic lumen. A virtual camera is used to create a set of images that provide a metric for matching pairs of folds between the prone and supine acquisitions. Image patches are generated at the fold positions using depth map renderings of the endoluminal surface and optimised by performing a virtual camera registration over a restricted set of degrees of freedom. The intensity difference between image pairs, along with additional neighbourhood information to enforce geometric constraints over a 2D parameterisation of the 3D space, are used as unary and pair-wise costs respectively, and included in a Markov Random Field (MRF) model to estimate the maximum a-posteriori fold labelling assignment.
The method achieved fold matching accuracy of 96.0% and 96.1% in patient cases with and without local colonic collapse. Moreover, it improved upon an existing surface-based registration algorithm by providing an initialisation. The set of landmark correspondences is used to non-rigidly transform a 2D source image derived from a conformal mapping process on the 3D endoluminal surface mesh. This achieves full surface correspondence between prone and supine views and can be further refined with an intensity based registration showing a statistically significant improvement (p<0.001p<0.001), and decreasing mean error from 11.9mm11.9mm to 6.0mm6.0mm measured at 1743 reference points from 17 CTC datasets
Deep Neural Networks for Anatomical Brain Segmentation
We present a novel approach to automatically segment magnetic resonance (MR)
images of the human brain into anatomical regions. Our methodology is based on
a deep artificial neural network that assigns each voxel in an MR image of the
brain to its corresponding anatomical region. The inputs of the network capture
information at different scales around the voxel of interest: 3D and orthogonal
2D intensity patches capture the local spatial context while large, compressed
2D orthogonal patches and distances to the regional centroids enforce global
spatial consistency. Contrary to commonly used segmentation methods, our
technique does not require any non-linear registration of the MR images. To
benchmark our model, we used the dataset provided for the MICCAI 2012 challenge
on multi-atlas labelling, which consists of 35 manually segmented MR images of
the brain. We obtained competitive results (mean dice coefficient 0.725, error
rate 0.163) showing the potential of our approach. To our knowledge, our
technique is the first to tackle the anatomical segmentation of the whole brain
using deep neural networks
Recommended from our members
Three-Dimensional Quantitative Assessment of Ablation Margins Based on Registration of Pre- and Post-Procedural MRI and Distance Map
Purpose: Contrast-enhanced MR images are widely used to confirm the adequacy of ablation margin after liver ablation for early prediction of local recurrence. However, quantitative assessment of the ablation margin by comparing pre- and post-procedural images remains challenging. We developed and tested a novel method for three-dimensional quantitative assessment of ablation margin based on non-rigid image registration and 3D distance map. Methods: Our method was tested with pre- and post-procedural MR images acquired in 21 patients who underwent image-guided percutaneous liver ablation. The two images were co-registered using non-rigid intensity-based registration. After the tumor and ablation volumes were segmented, target volume coverage, percent of tumor coverage, and Dice Similarity Coefficient were calculated as metrics representing overall adequacy of ablation. In addition, 3D distance map around the tumor was computed and superimposed on the ablation volume to identify the area with insufficient margins. For patients with local recurrences, the follow-up images were registered to the post-procedural image. Three-D minimum distance between the recurrence and the areas with insufficient margins were quantified. Results: The percent tumor coverage for all non-recurrent cases was 100%. Five cases had tumor recurrences, and the 3D distance map revealed insufficient tumor coverage or a 0-millimeter margin. It also showed that two recurrences were remote to the insufficient margin. Conclusions: Non-rigid registration and 3D distance map allows us to quantitatively evaluate the adequacy of the ablation margin after percutaneous liver ablation. The method may be useful to predict local recurrences immediately following ablation procedure
Registration of Standardized Histological Images in Feature Space
In this paper, we propose three novel and important methods for the
registration of histological images for 3D reconstruction. First, possible
intensity variations and nonstandardness in images are corrected by an
intensity standardization process which maps the image scale into a standard
scale where the similar intensities correspond to similar tissues meaning.
Second, 2D histological images are mapped into a feature space where continuous
variables are used as high confidence image features for accurate registration.
Third, we propose an automatic best reference slice selection algorithm that
improves reconstruction quality based on both image entropy and mean square
error of the registration process. We demonstrate that the choice of reference
slice has a significant impact on registration error, standardization, feature
space and entropy information. After 2D histological slices are registered
through an affine transformation with respect to an automatically chosen
reference, the 3D volume is reconstructed by co-registering 2D slices
elastically.Comment: SPIE Medical Imaging 2008 - submissio
- …