7,149 research outputs found
Dual-Stream Pyramid Registration Network
We propose a Dual-Stream Pyramid Registration Network (referred as
Dual-PRNet) for unsupervised 3D medical image registration. Unlike recent
CNN-based registration approaches, such as VoxelMorph, which explores a
single-stream encoder-decoder network to compute a registration fields from a
pair of 3D volumes, we design a two-stream architecture able to compute
multi-scale registration fields from convolutional feature pyramids. Our
contributions are two-fold: (i) we design a two-stream 3D encoder-decoder
network which computes two convolutional feature pyramids separately for a pair
of input volumes, resulting in strong deep representations that are meaningful
for deformation estimation; (ii) we propose a pyramid registration module able
to predict multi-scale registration fields directly from the decoding feature
pyramids. This allows it to refine the registration fields gradually in a
coarse-to-fine manner via sequential warping, and enable the model with the
capability for handling significant deformations between two volumes, such as
large displacements in spatial domain or slice space. The proposed Dual-PRNet
is evaluated on two standard benchmarks for brain MRI registration, where it
outperforms the state-of-the-art approaches by a large margin, e.g., having
improvements over recent VoxelMorph [2] with 0.683->0.778 on the LPBA40, and
0.511->0.631 on the Mindboggle101, in term of average Dice score.Comment: To appear in MICCAI 2019 (Oral
Improving Misaligned Multi-modality Image Fusion with One-stage Progressive Dense Registration
Misalignments between multi-modality images pose challenges in image fusion,
manifesting as structural distortions and edge ghosts. Existing efforts
commonly resort to registering first and fusing later, typically employing two
cascaded stages for registration,i.e., coarse registration and fine
registration. Both stages directly estimate the respective target deformation
fields. In this paper, we argue that the separated two-stage registration is
not compact, and the direct estimation of the target deformation fields is not
accurate enough. To address these challenges, we propose a Cross-modality
Multi-scale Progressive Dense Registration (C-MPDR) scheme, which accomplishes
the coarse-to-fine registration exclusively using a one-stage optimization,
thus improving the fusion performance of misaligned multi-modality images.
Specifically, two pivotal components are involved, a dense Deformation Field
Fusion (DFF) module and a Progressive Feature Fine (PFF) module. The DFF
aggregates the predicted multi-scale deformation sub-fields at the current
scale, while the PFF progressively refines the remaining misaligned features.
Both work together to accurately estimate the final deformation fields. In
addition, we develop a Transformer-Conv-based Fusion (TCF) subnetwork that
considers local and long-range feature dependencies, allowing us to capture
more informative features from the registered infrared and visible images for
the generation of high-quality fused images. Extensive experimental analysis
demonstrates the superiority of the proposed method in the fusion of misaligned
cross-modality images
Non-iterative Coarse-to-fine Transformer Networks for Joint Affine and Deformable Image Registration
Image registration is a fundamental requirement for medical image analysis.
Deep registration methods based on deep learning have been widely recognized
for their capabilities to perform fast end-to-end registration. Many deep
registration methods achieved state-of-the-art performance by performing
coarse-to-fine registration, where multiple registration steps were iterated
with cascaded networks. Recently, Non-Iterative Coarse-to-finE (NICE)
registration methods have been proposed to perform coarse-to-fine registration
in a single network and showed advantages in both registration accuracy and
runtime. However, existing NICE registration methods mainly focus on deformable
registration, while affine registration, a common prerequisite, is still
reliant on time-consuming traditional optimization-based methods or extra
affine registration networks. In addition, existing NICE registration methods
are limited by the intrinsic locality of convolution operations. Transformers
may address this limitation for their capabilities to capture long-range
dependency, but the benefits of using transformers for NICE registration have
not been explored. In this study, we propose a Non-Iterative Coarse-to-finE
Transformer network (NICE-Trans) for image registration. Our NICE-Trans is the
first deep registration method that (i) performs joint affine and deformable
coarse-to-fine registration within a single network, and (ii) embeds
transformers into a NICE registration framework to model long-range relevance
between images. Extensive experiments with seven public datasets show that our
NICE-Trans outperforms state-of-the-art registration methods on both
registration accuracy and runtime.Comment: Accepted at International Conference on Medical Image Computing and
Computer Assisted Intervention (MICCAI 2023
- …