2 research outputs found
Plug-and-Play Priors for Reconstruction-based Placental Image Registration
This paper presents a novel deformable registration framework, leveraging an
image prior specified through a denoising function, for severely
noise-corrupted placental images. Recent work on plug-and-play (PnP) priors has
shown the state-of-the-art performance of reconstruction algorithms under such
priors in a range of imaging applications. Integration of powerful image
denoisers into advanced registration methods provides our model with a
flexibility to accommodate datasets that have low signal-to-noise ratios
(SNRs). We demonstrate the performance of our method under a wide variety of
denoising models in the context of diffeomorphic image registration.
Experimental results show that our model substantially improves the accuracy of
spatial alignment in applications of 3D in-utero diffusion-weighted MR images
(DW-MRI) that suffer from low SNR and large spatial transformations
Deep Learning for Regularization Prediction in Diffeomorphic Image Registration
This paper presents a predictive model for estimating regularization
parameters of diffeomorphic image registration. We introduce a novel framework
that automatically determines the parameters controlling the smoothness of
diffeomorphic transformations. Our method significantly reduces the effort of
parameter tuning, which is time and labor-consuming. To achieve the goal, we
develop a predictive model based on deep convolutional neural networks (CNN)
that learns the mapping between pairwise images and the regularization
parameter of image registration. In contrast to previous methods that estimate
such parameters in a high-dimensional image space, our model is built in an
efficient bandlimited space with much lower dimensions. We demonstrate the
effectiveness of our model on both 2D synthetic data and 3D real brain images.
Experimental results show that our model not only predicts appropriate
regularization parameters for image registration, but also improving the
network training in terms of time and memory efficiency.Comment: 20 pages, 8 figure