87 research outputs found
Generation and Recombination for Multifocus Image Fusion with Free Number of Inputs
Multifocus image fusion is an effective way to overcome the limitation of
optical lenses. Many existing methods obtain fused results by generating
decision maps. However, such methods often assume that the focused areas of the
two source images are complementary, making it impossible to achieve
simultaneous fusion of multiple images. Additionally, the existing methods
ignore the impact of hard pixels on fusion performance, limiting the visual
quality improvement of fusion image. To address these issues, a combining
generation and recombination model, termed as GRFusion, is proposed. In
GRFusion, focus property detection of each source image can be implemented
independently, enabling simultaneous fusion of multiple source images and
avoiding information loss caused by alternating fusion. This makes GRFusion
free from the number of inputs. To distinguish the hard pixels from the source
images, we achieve the determination of hard pixels by considering the
inconsistency among the detection results of focus areas in source images.
Furthermore, a multi-directional gradient embedding method for generating full
focus images is proposed. Subsequently, a hard-pixel-guided recombination
mechanism for constructing fused result is devised, effectively integrating the
complementary advantages of feature reconstruction-based method and focused
pixel recombination-based method. Extensive experimental results demonstrate
the effectiveness and the superiority of the proposed method.The source code
will be released on https://github.com/xxx/xxx
Spatial Stimuli Gradient Based Multifocus Image Fusion Using Multiple Sized Kernels
Multi-focus image fusion technique extracts the focused areas from all the source images and combines them into a new image which contains all focused objects. This paper proposes a spatial domain fusion scheme for multi-focus images by using multiple size kernels. Firstly, source images are pre-processed with a contrast enhancement step and then the soft and hard decision maps are generated by employing a sliding window technique using multiple sized kernels on the gradient images. Hard decision map selects the accurate focus information from the source images, whereas, the soft decision map selects the basic focus information and contains minimum falsely detected focused/unfocused regions. These decision maps are further processed to compute the final focus map. Gradient images are constructed through state-ofthe-art edge detection technique, spatial stimuli gradient sketch model, which computes the local stimuli from perceived brightness and hence enhances the essential structural and edge information. Detailed experiment results demonstrate that the proposed multi-focus image fusion algorithm performs better than the other well known state-of-the-art multifocus image fusion methods, in terms of subjective visual perception and objective quality evaluation metrics
Fast filtering image fusion
© 2017 SPIE and IS & T. Image fusion aims at exploiting complementary information in multimodal images to create a single composite image with extended information content. An image fusion framework is proposed for different types of multimodal images with fast filtering in the spatial domain. First, image gradient magnitude is used to detect contrast and image sharpness. Second, a fast morphological closing operation is performed on image gradient magnitude to bridge gaps and fill holes. Third, the weight map is obtained from the multimodal image gradient magnitude and is filtered by a fast structure-preserving filter. Finally, the fused image is composed by using a weighed-sum rule. Experimental results on several groups of images show that the proposed fast fusion method has a better performance than the state-of-the-art methods, running up to four times faster than the fastest baseline algorithm
Structural similarity loss for learning to fuse multi-focus images
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. Convolutional neural networks have recently been used for multi-focus image fusion. However, some existing methods have resorted to adding Gaussian blur to focused images, to simulate defocus, thereby generating data (with ground-truth) for supervised learning. Moreover, they classify pixels as ‘focused’ or ‘defocused’, and use the classified results to construct the fusion weight maps. This then necessitates a series of post-processing steps. In this paper, we present an end-to-end learning approach for directly predicting the fully focused output image from multi-focus input image pairs. The suggested approach uses a CNN architecture trained to perform fusion, without the need for ground truth fused images. The CNN exploits the image structural similarity (SSIM) to calculate the loss, a metric that is widely accepted for fused image quality evaluation. What is more, we also use the standard deviation of a local window of the image to automatically estimate the importance of the source images in the final fused image when designing the loss function. Our network can accept images of variable sizes and hence, we are able to utilize real benchmark datasets, instead of simulated ones, to train our network. The model is a feed-forward, fully convolutional neural network that can process images of variable sizes during test time. Extensive evaluation on benchmark datasets show that our method outperforms, or is comparable with, existing state-of-the-art techniques on both objective and subjective benchmarks
- …