6,790 research outputs found
Combined Mutual Information of Intensity and Gradient for Multi-modal Medical Image Registration
In this thesis, registration methods for multi-modal medical images are reviewed with mutual information-based methods discussed in detail. Since it was proposed, mutual information has gained intensive research and is getting very popular, however its robustness is questionable and may fail in some cases. The possible reason might be it does not consider the spatial information in the image pair. In order to improve this measure, the thesis proposes to use combined mutual information of intensity and gradient for multi-modal medical image registration. The proposed measure utilizes both the intensity and gradient information of an image pair. Maximization of this measure is assumed to correctly register an image pair. Optimization of the registration measure in a multi-dimensional space is another major issue in multi-modal medical image registration. The thesis first briefly reviews the commonly used optimization techniques and then discusses in detail the Powell\u27s conjugate direction set method, which is implemented to find the maximum of the combined mutual information of an image pair. In the experiment, we first register slice images scanned in a single patient in the same or different scanning sessions by the proposed method. Then 20 pairs of co-registered CT and PET slice images at three different resolutions are used to study the performance of the proposed measure and four other measures discussed in this thesis. Experimental results indicate that the proposed combined measure produces reliable registrations and it outperforms the intensity- and gradient-based measures at all three resolutions
Methods for multi-spectral image fusion: identifying stable and repeatable information across the visible and infrared spectra
Fusion of images captured from different viewpoints is a well-known challenge in computer vision with many established approaches and applications; however, if the observations are captured by sensors also separated by wavelength, this challenge is compounded significantly. This dissertation presents an investigation into the fusion of visible and thermal image information from two front-facing sensors mounted side-by-side. The primary focus of this work is the development of methods that enable us to map and overlay multi-spectral information; the goal is to establish a combined image in which each pixel contains both colour and thermal information. Pixel-level fusion of these distinct modalities is approached using computational stereo methods; the focus is on the viewpoint alignment and correspondence search/matching stages of processing. Frequency domain analysis is performed using a method called phase congruency. An extensive investigation of this method is carried out with two major objectives: to identify predictable relationships between the elements extracted from each modality, and to establish a stable representation of the common information captured by both sensors. Phase congruency is shown to be a stable edge detector and repeatable spatial similarity measure for multi-spectral information; this result forms the basis for the methods developed in the subsequent chapters of this work. The feasibility of automatic alignment with sparse feature-correspondence methods is investigated. It is found that conventional methods fail to match inter-spectrum correspondences, motivating the development of an edge orientation histogram (EOH) descriptor which incorporates elements of the phase congruency process. A cost function, which incorporates the outputs of the phase congruency process and the mutual information similarity measure, is developed for computational stereo correspondence matching. An evaluation of the proposed cost function shows it to be an effective similarity measure for multi-spectral information
Learning Deep Similarity Metric for 3D MR-TRUS Registration
Purpose: The fusion of transrectal ultrasound (TRUS) and magnetic resonance
(MR) images for guiding targeted prostate biopsy has significantly improved the
biopsy yield of aggressive cancers. A key component of MR-TRUS fusion is image
registration. However, it is very challenging to obtain a robust automatic
MR-TRUS registration due to the large appearance difference between the two
imaging modalities. The work presented in this paper aims to tackle this
problem by addressing two challenges: (i) the definition of a suitable
similarity metric and (ii) the determination of a suitable optimization
strategy.
Methods: This work proposes the use of a deep convolutional neural network to
learn a similarity metric for MR-TRUS registration. We also use a composite
optimization strategy that explores the solution space in order to search for a
suitable initialization for the second-order optimization of the learned
metric. Further, a multi-pass approach is used in order to smooth the metric
for optimization.
Results: The learned similarity metric outperforms the classical mutual
information and also the state-of-the-art MIND feature based methods. The
results indicate that the overall registration framework has a large capture
range. The proposed deep similarity metric based approach obtained a mean TRE
of 3.86mm (with an initial TRE of 16mm) for this challenging problem.
Conclusion: A similarity metric that is learned using a deep neural network
can be used to assess the quality of any given image registration and can be
used in conjunction with the aforementioned optimization framework to perform
automatic registration that is robust to poor initialization.Comment: To appear on IJCAR
Robust and Fast 3D Scan Alignment using Mutual Information
This paper presents a mutual information (MI) based algorithm for the
estimation of full 6-degree-of-freedom (DOF) rigid body transformation between
two overlapping point clouds. We first divide the scene into a 3D voxel grid
and define simple to compute features for each voxel in the scan. The two scans
that need to be aligned are considered as a collection of these features and
the MI between these voxelized features is maximized to obtain the correct
alignment of scans. We have implemented our method with various simple point
cloud features (such as number of points in voxel, variance of z-height in
voxel) and compared the performance of the proposed method with existing
point-to-point and point-to- distribution registration methods. We show that
our approach has an efficient and fast parallel implementation on GPU, and
evaluate the robustness and speed of the proposed algorithm on two real-world
datasets which have variety of dynamic scenes from different environments
Geometric and photometric affine invariant image registration
This thesis aims to present a solution to the correspondence problem for the registration
of wide-baseline images taken from uncalibrated cameras. We propose an affine
invariant descriptor that combines the geometry and photometry of the scene to find
correspondences between both views. The geometric affine invariant component of the
descriptor is based on the affine arc-length metric, whereas the photometry is analysed
by invariant colour moments. A graph structure represents the spatial distribution of the
primitive features; i.e. nodes correspond to detected high-curvature points, whereas arcs
represent connectivities by extracted contours. After matching, we refine the search for
correspondences by using a maximum likelihood robust algorithm. We have evaluated
the system over synthetic and real data. The method is endemic to propagation of errors
introduced by approximations in the system.BAE SystemsSelex Sensors and Airborne System
- …