25 research outputs found
Fusion of Infrared and Visible Images Based on Non-subsample Contourlet Transform
For the single spectrum image could not fully express the target feature information, this paper proposed a multispectral image fusion method based on non-subsample contourlet transform (NSCT). For the low frequency coefficients decomposed, fourth-order correlation coefficient is used to calculate the correlation between each low frequency coefficients, averaging fusion for the higher correlation coefficient, weight phase congruency fusion for the low correlation coefficient. For high frequency coefficients, Gaussian weight sum modified Laplace method is used for fusing, to retain more local structure details. Simulation results show that the method effectively retain the image structure information and more local details, and increase the image contrast
Construction of all-in-focus images assisted by depth sensing
Multi-focus image fusion is a technique for obtaining an all-in-focus image
in which all objects are in focus to extend the limited depth of field (DoF) of
an imaging system. Different from traditional RGB-based methods, this paper
presents a new multi-focus image fusion method assisted by depth sensing. In
this work, a depth sensor is used together with a color camera to capture
images of a scene. A graph-based segmentation algorithm is used to segment the
depth map from the depth sensor, and the segmented regions are used to guide a
focus algorithm to locate in-focus image blocks from among multi-focus source
images to construct the reference all-in-focus image. Five test scenes and six
evaluation metrics were used to compare the proposed method and representative
state-of-the-art algorithms. Experimental results quantitatively demonstrate
that this method outperforms existing methods in both speed and quality (in
terms of comprehensive fusion metrics). The generated images can potentially be
used as reference all-in-focus images.Comment: 18 pages. This paper has been submitted to Computer Vision and Image
Understandin
A Novel Region based Image Fusion Method using Highboost Filtering and Fuzzy Logic
This paper proposes a novel region based image
fusion scheme based on high boost filtering concept using discrete
wavelet transform. In the recent literature, region based image
fusion methods show better performance than pixel based image
fusion method. Proposed method is a novel idea which uses
high boost filtering concept to get an accurate segmentation
using discrete wavelet transform. This concept is used to extract
regions form input registered source images which is than
compared with different fusion rules. The fusion rule based on
spatial frequency and standard deviation is also proposed to fuse
multimodality images. The different fusion rules are applied on
various categories of input source images and resultant fused
image is generated. Proposed method is applied on registered
images of multifocus and multimodality images and results are
compared using standard reference based and non-reference
based image fusion parameters. It has been observed from
simulation results that our proposed algorithm is consistent and
preserves more information compared to earlier reported pixel
based and region based methods
The development of the quaternion wavelet transform
The purpose of this article is to review what has been written on what other authors have called quaternion wavelet transforms (QWTs): there is no consensus about what these should look like and what their properties should be. We briefly explain what real continuous and discrete wavelet transforms and multiresolution analysis are and why complex wavelet transforms were introduced; we then go on to detail published approaches to QWTs and to analyse them. We conclude with our own analysis of what it is that should define a QWT as being truly quaternionic and why all but a few of the “QWTs” we have described do not fit our definition
Region based Multimodality Image Fusion Method
This paper proposes a novel region based image
fusion scheme based on high boost filtering concept using discrete
wavelet transform. In the recent literature, region based image
fusion methods show better performance than pixel based image
fusion method. The graph based normalized cutest algorithm
is used for image segmentation. Proposed method is a novel
idea which uses high boost filtering concept to get an accurate
segmentation using discrete wavelet transform. This concept is
used to extract regions from input registered source images which
is then compared with different fusion rules. The new MMS
fusion rule is also proposed to fuse multimodality images. The
different fusion rules are applied on various categories of input
source images and resultant fused image is generated. Proposed
method is applied on large number of registered images of various
categories of multifocus and multimodality images and results
are compared using standard reference based and nonreference
based image fusion parameters. It has been observed from
simulation results that our proposed algorithm is consistent and
preserves more information compared to earlier reported pixel
based and region based methods
NSCT Based Multimodal Fusion Technique for Medical Images
In this paper my idea is to propose a new approach for multimodal medical image fusion based on NSCT which further gone ease the work for medical images .Multimodality in medical imaging are X-ray, computed tomography (CT), magnetic resonance imaging (MRI), magnetic resonance angiography (MRA), and positron emission tomography (PET). The reason behind writing this paper is to let researcher get acquainted with idea of multimodal image using a technique called as non-sampled contourlet transform (NSCT) by the help of this technique we can capture all relevant information required for medical diagnosis non-sub sampled contourlet transform (NSCT). The two multimodality medical images are first transformed by NSCT into low- and high-frequency components followed by combining the low- and high-frequency components. Phase congruency and directive contrast are main methods which are proposed for various need of low frequency and high frequency coefficients .Finally NSCT Based method is used for medical multimodal images
Non-Standard Imaging Techniques
The first objective of the thesis is to investigate the problem of reconstructing a small-scale object (a few millimeters or smaller) in 3D. In Chapter 3, we show how this problem can be solved effectively by a new multifocus multiview 3D reconstruction procedure which includes a new Fixed-Lens multifocus image capture and a calibrated image registration technique using analytic homography transformation. The experimental results using the real and synthetic images demonstrate the effectiveness of the proposed solutions by showing that both the fixed-lens image capture and multifocus stacking with calibrated image alignment significantly reduce the errors in the camera poses and produce more complete 3D reconstructed models as compared with those by the conventional moving lens image capture and multifocus stacking.
The second objective of the thesis is modelling the dual-pixel (DP) camera. In Chapter 4, to understand the potential of the DP sensor for computer vision applications, we study the formation of the DP pair which links the blur and the depth information. A mathematical DP model is proposed which can benefit depth estimation by the blur. These explorations motivate us to propose an end-to-end DDDNet (DP-based Depth and Deblur Network) to jointly estimate the depth and restore the image . Moreover, we define a reblur loss, which reflects the relationship of the DP image formation process with depth information, to regularize our depth estimate in training. To meet the requirement of a large amount of data for learning, we propose the first DP image simulator which allows us to create datasets with DP pairs from any existing RGBD dataset. As a side contribution, we collect a real dataset for further research. Extensive experimental evaluation on both synthetic and real datasets shows that our approach achieves competitive performance compared to state-of-the-art approaches.
Another (third) objective of this thesis is to tackle the multifocus image fusion problem, particularly for long multifocus image sequences. Multifocus image stacking/fusion produces an in-focus image of a scene from a number of partially focused images of that scene in order to extend the depth of field. One of the limitations of the current state of the art multifocus fusion methods is not considering image registration/alignment before fusion. Consequently, fusing unregistered multifocus images produces an in-focus image containing misalignment artefacts. In Chapter 5, we propose image registration by projective transformation before fusion to remove the misalignment artefacts. We also propose a method based on 3D deconvolution to retrieve the in-focus image by formulating the multifocus image fusion problem as a 3D deconvolution problem. The proposed method achieves superior performance compared to the state of the art methods. It is also shown that, the proposed projective transformation for image registration can improve the quality of the fused images. Moreover, we implement a multifocus simulator to generate synthetic multifocus data from any RGB-D dataset. The fourth objective of this thesis is to explore new ways to detect the polarization state of light. To achieve the objective, in Chapter 6, we investigate a new optical filter namely optical rotation filter for detecting the polarization state with a fewer number of images. The proposed method can estimate polarization state using two images, one with the filter and another without. The accuracy of estimating the polarization parameters using the proposed method is almost similar to that of the existing state of the art method. In addition, the feasibility of detecting the polarization state using only one RGB image captured with the optical rotation filter is also demonstrated by estimating the image without the filter from the image with the filter using a generative adversarial network
An Improved Infrared/Visible Fusion for Astronomical Images
An undecimated dual tree complex wavelet transform (UDTCWT) based fusion scheme for astronomical visible/IR images is developed. The UDTCWT reduces noise effects and improves object classification due to its inherited shift invariance property. Local standard deviation and distance transforms are used to extract useful information (especially small objects). Simulation results compared with the state-of-the-art fusion techniques illustrate the superiority of proposed scheme in terms of accuracy for most of the cases