19 research outputs found
A Novel Fusion Framework Based on Adaptive PCNN in NSCT Domain for Whole-Body PET and CT Images
The PET and CT fusion images, combining the anatomical and functional information, have important clinical meaning. This paper proposes a novel fusion framework based on adaptive pulse-coupled neural networks (PCNNs) in nonsubsampled contourlet transform (NSCT) domain for fusing whole-body PET and CT images. Firstly, the gradient average of each pixel is chosen as the linking strength of PCNN model to implement self-adaptability. Secondly, to improve the fusion performance, the novel sum-modified Laplacian (NSML) and energy of edge (EOE) are extracted as the external inputs of the PCNN models for low- and high-pass subbands, respectively. Lastly, the rule of max region energy is adopted as the fusion rule and different energy templates are employed in the low- and high-pass subbands. The experimental results on whole-body PET and CT data (239 slices contained by each modality) show that the proposed framework outperforms the other six methods in terms of the seven commonly used fusion performance metrics
Comparative Analysis and Fusion of MRI and PET Images based on Wavelets for Clinical Diagnosis
Nowadays, Medical imaging modalities like Magnetic Resonance Imaging (MRI), Positron Emission Tomography (PET), Single Photon Emission Tomography (SPECT), and Computed Tomography (CT) play a crucial role in clinical diagnosis and treatment planning. The images obtained from each of these modalities contain complementary information of the organ imaged. Image fusion algorithms are employed to bring all of this disparate information together into a single image, allowing doctors to diagnose disorders quickly. This paper proposes a novel technique for the fusion of MRI and PET images based on YUV color space and wavelet transform. Quality assessment based on entropy showed that the method can achieve promising results for medical image fusion. The paper has done a comparative analysis of the fusion of MRI and PET images using different wavelet families at various decomposition levels for the detection of brain tumors as well as Alzheimer’s disease. The quality assessment and visual analysis showed that the Dmey wavelet at decomposition level 3 is optimum for the fusion of MRI and PET images. This paper also compared the results of several fusion rules such as average, maximum, and minimum, finding that the maximum fusion rule outperformed the other two
An Attention-based Multi-Scale Feature Learning Network for Multimodal Medical Image Fusion
Medical images play an important role in clinical applications. Multimodal
medical images could provide rich information about patients for physicians to
diagnose. The image fusion technique is able to synthesize complementary
information from multimodal images into a single image. This technique will
prevent radiologists switch back and forth between different images and save
lots of time in the diagnostic process. In this paper, we introduce a novel
Dilated Residual Attention Network for the medical image fusion task. Our
network is capable to extract multi-scale deep semantic features. Furthermore,
we propose a novel fixed fusion strategy termed Softmax-based weighted strategy
based on the Softmax weights and matrix nuclear norm. Extensive experiments
show our proposed network and fusion strategy exceed the state-of-the-art
performance compared with reference image fusion methods on four commonly used
fusion metrics.Comment: 8 pages, 8 figures, 3 table
Multifocus Image Fusion Using Biogeography-Based Optimization
For multifocus image fusion in spatial domain, sharper blocks from different source images are selected to fuse a new image. Block size significantly affects the fusion results and a fixed block size is not applicable in various multifocus images. In this paper, a novel multifocus image fusion algorithm using biogeography-based optimization is proposed to obtain the optimal block size. The sharper blocks of each source image are first selected by sum modified Laplacian and morphological filter to contain an initial fused image. Then, the proposed algorithm uses the migration and mutation operation of biogeography-based optimization to search the optimal block size according to the fitness function in respect of spatial frequency. The chaotic search is adopted during iteration to improve optimization precision. The final fused image is constructed based on the optimal block size. Experimental results demonstrate that the proposed algorithm has good quantitative and visual evaluations
MDLatLRR: A novel decomposition method for infrared and visible image fusion
Image decomposition is crucial for many image processing tasks, as it allows
to extract salient features from source images. A good image decomposition
method could lead to a better performance, especially in image fusion tasks. We
propose a multi-level image decomposition method based on latent low-rank
representation(LatLRR), which is called MDLatLRR. This decomposition method is
applicable to many image processing fields. In this paper, we focus on the
image fusion task. We develop a novel image fusion framework based on MDLatLRR,
which is used to decompose source images into detail parts(salient features)
and base parts. A nuclear-norm based fusion strategy is used to fuse the detail
parts, and the base parts are fused by an averaging strategy. Compared with
other state-of-the-art fusion methods, the proposed algorithm exhibits better
fusion performance in both subjective and objective evaluation.Comment: IEEE Trans. Image Processing 2020, 14 pages, 17 figures, 3 table
A multimodal fusion method for Alzheimer’s disease based on DCT convolutional sparse representation
IntroductionThe medical information contained in magnetic resonance imaging (MRI) and positron emission tomography (PET) has driven the development of intelligent diagnosis of Alzheimer’s disease (AD) and multimodal medical imaging. To solve the problems of severe energy loss, low contrast of fused images and spatial inconsistency in the traditional multimodal medical image fusion methods based on sparse representation. A multimodal fusion algorithm for Alzheimer’ s disease based on the discrete cosine transform (DCT) convolutional sparse representation is proposed.MethodsThe algorithm first performs a multi-scale DCT decomposition of the source medical images and uses the sub-images of different scales as training images, respectively. Different sparse coefficients are obtained by optimally solving the sub-dictionaries at different scales using alternating directional multiplication method (ADMM). Secondly, the coefficients of high-frequency and low-frequency subimages are inverse DCTed using an improved L1 parametric rule combined with improved spatial frequency novel sum-modified SF (NMSF) to obtain the final fused images.Results and discussionThrough extensive experimental results, we show that our proposed method has good performance in contrast enhancement, texture and contour information retention