7,225 research outputs found
Registration of Standardized Histological Images in Feature Space
In this paper, we propose three novel and important methods for the
registration of histological images for 3D reconstruction. First, possible
intensity variations and nonstandardness in images are corrected by an
intensity standardization process which maps the image scale into a standard
scale where the similar intensities correspond to similar tissues meaning.
Second, 2D histological images are mapped into a feature space where continuous
variables are used as high confidence image features for accurate registration.
Third, we propose an automatic best reference slice selection algorithm that
improves reconstruction quality based on both image entropy and mean square
error of the registration process. We demonstrate that the choice of reference
slice has a significant impact on registration error, standardization, feature
space and entropy information. After 2D histological slices are registered
through an affine transformation with respect to an automatically chosen
reference, the 3D volume is reconstructed by co-registering 2D slices
elastically.Comment: SPIE Medical Imaging 2008 - submissio
To Learn or Not to Learn Features for Deformable Registration?
Feature-based registration has been popular with a variety of features
ranging from voxel intensity to Self-Similarity Context (SSC). In this paper,
we examine the question on how features learnt using various Deep Learning (DL)
frameworks can be used for deformable registration and whether this feature
learning is necessary or not. We investigate the use of features learned by
different DL methods in the current state-of-the-art discrete registration
framework and analyze its performance on 2 publicly available datasets. We draw
insights into the type of DL framework useful for feature learning and the
impact, if any, of the complexity of different DL models and brain parcellation
methods on the performance of discrete registration. Our results indicate that
the registration performance with DL features and SSC are comparable and stable
across datasets whereas this does not hold for low level features.Comment: 9 pages, 4 figure
An Unsupervised Learning Model for Deformable Medical Image Registration
We present a fast learning-based algorithm for deformable, pairwise 3D
medical image registration. Current registration methods optimize an objective
function independently for each pair of images, which can be time-consuming for
large data. We define registration as a parametric function, and optimize its
parameters given a set of images from a collection of interest. Given a new
pair of scans, we can quickly compute a registration field by directly
evaluating the function using the learned parameters. We model this function
using a convolutional neural network (CNN), and use a spatial transform layer
to reconstruct one image from another while imposing smoothness constraints on
the registration field. The proposed method does not require supervised
information such as ground truth registration fields or anatomical landmarks.
We demonstrate registration accuracy comparable to state-of-the-art 3D image
registration, while operating orders of magnitude faster in practice. Our
method promises to significantly speed up medical image analysis and processing
pipelines, while facilitating novel directions in learning-based registration
and its applications. Our code is available at
https://github.com/balakg/voxelmorph .Comment: 9 pages, in CVPR 201
A Deep Learning Approach to Denoise Optical Coherence Tomography Images of the Optic Nerve Head
Purpose: To develop a deep learning approach to de-noise optical coherence
tomography (OCT) B-scans of the optic nerve head (ONH).
Methods: Volume scans consisting of 97 horizontal B-scans were acquired
through the center of the ONH using a commercial OCT device (Spectralis) for
both eyes of 20 subjects. For each eye, single-frame (without signal
averaging), and multi-frame (75x signal averaging) volume scans were obtained.
A custom deep learning network was then designed and trained with 2,328 "clean
B-scans" (multi-frame B-scans), and their corresponding "noisy B-scans" (clean
B-scans + gaussian noise) to de-noise the single-frame B-scans. The performance
of the de-noising algorithm was assessed qualitatively, and quantitatively on
1,552 B-scans using the signal to noise ratio (SNR), contrast to noise ratio
(CNR), and mean structural similarity index metrics (MSSIM).
Results: The proposed algorithm successfully denoised unseen single-frame OCT
B-scans. The denoised B-scans were qualitatively similar to their corresponding
multi-frame B-scans, with enhanced visibility of the ONH tissues. The mean SNR
increased from dB (single-frame) to dB
(denoised). For all the ONH tissues, the mean CNR increased from (single-frame) to (denoised). The MSSIM increased from
(single frame) to (denoised) when compared with
the corresponding multi-frame B-scans.
Conclusions: Our deep learning algorithm can denoise a single-frame OCT
B-scan of the ONH in under 20 ms, thus offering a framework to obtain superior
quality OCT B-scans with reduced scanning times and minimal patient discomfort
- …