18 research outputs found
Disparity estimation using TI multi-wavelet transform
A multi-resolution image matching technique based on translation invariant discrete multi-wavelet transform followed by a coarse to fine matching strategy is presented. The technique addresses the estimation of optimal corresponding points and the corresponding disparity maps in the presence of occlusion, ambiguity and illuminative variations in the two perspective views taken by two different cameras or at different lighting conditions. The problem of occlusion and ambiguity is addressed explicitly by a geometric optimization approach along with the uniqueness constraint whereas the illuminative variation is dealt with by using windowed normalized correlation on the discrete multi-wavelet coefficients.<br /
Efficient Stereo Matching with Decoupled Dissimilarity Measure Using Successive Weighted Summation
Developing matching algorithms from stereo image pairs to obtain correct disparity maps for 3D reconstruction has been the focus of intensive research. A constant computational complexity algorithm to calculate dissimilarity aggregation in assessing disparity based on separable successive weighted summation (SWS) among horizontal and vertical directions was proposed but still not satisfactory. This paper presents a novel method which enables decoupled dissimilarity measure in the aggregation, further improving the accuracy and robustness of stereo correspondence. The aggregated cost is also used to refine disparities based on a local curve-fitting procedure. According to our experimental results on Middlebury benchmark evaluation, the proposed approach has comparable performance when compared with the selected state-of-the-art algorithms and has the lowest mismatch rate. Besides, the refinement procedure is shown to be capable of preserving object boundaries and depth discontinuities while smoothing out disparity maps
3D RECONSTRUCTION FROM STEREO/RANGE IMAGES
3D reconstruction from stereo/range image is one of the most fundamental and extensively researched topics in computer vision. Stereo research has recently experienced somewhat of a new era, as a result of publically available performance testing such as the Middlebury data set, which has allowed researchers to compare their algorithms against all the state-of-the-art algorithms. This thesis investigates into the general stereo problems in both the two-view stereo and multi-view stereo scopes. In the two-view stereo scope, we formulate an algorithm for the stereo matching problem with careful handling of disparity, discontinuity and occlusion. The algorithm works with a global matching stereo model based on an energy minimization framework. The experimental results are evaluated on the Middlebury data set, showing that our algorithm is the top performer. A GPU approach of the Hierarchical BP algorithm is then proposed, which provides similar stereo quality to CPU Hierarchical BP while running at real-time speed. A fast-converging BP is also proposed to solve the slow convergence problem of general BP algorithms. Besides two-view stereo, ecient multi-view stereo for large scale urban reconstruction is carefully studied in this thesis. A novel approach for computing depth maps given urban imagery where often large parts of surfaces are weakly textured is presented. Finally, a new post-processing step to enhance the range images in both the both the spatial resolution and depth precision is proposed
LEVEL-BASED CORRESPONDENCE APPROACH TO COMPUTATIONAL STEREO
One fundamental problem in computational stereo reconstruction is correspondence.
Correspondence is the method of detecting the real world object reflections in two
camera views. This research focuses on correspondence, proposing an algorithm to
improve such detection for low quality cameras (webcams) while trying to achieve
real-time image processing.
Correspondence plays an important role in computational stereo reconstruction and it
has a vast spectrum of applicability. This method is useful in other areas such as
structure from motion reconstruction, object detection, tracking in robot vision and
virtual reality. Due to its importance, a correspondence method needs to be accurate
enough to meet the requirement of such fields but it should be less costly and easy to
use and configure, to be accessible by everyone.
By comparing current local correspondence method and discussing their weakness
and strength, this research tries to enhance an algorithm to improve previous works to
achieve fast detection, less costly and acceptable accuracy to meet the requirement of
reconstruction. In this research, the correspondence is divided into four stages. Two
stages of preprocessing which are noise reduction and edge detection have been
compared with respect to different methods available. In the next stage, the feature
detection process is introduced and discussed focusing on possible solutions to reduce
errors created by system or problem occurring in the scene such as occlusion. Lastly,
in the final stage it elaborates different methods of displaying reconstructed result.
Different sets of data are processed based on the steps involved in correspondence and
the results are discussed and compared in detail. The finding shows how this system
can achieve high speed and acceptable outcome despite of poor quality input. As a
conclusion, some possible improvements are proposed based on ultimate outcome
Recommended from our members
View synthesis for depth from motion 3D x-ray imaging.
The depth from motion or kinetic depth X-ray imaging (KDEX) technique is designed to enhance the luggage screening at airport checkpoints. The technique requires multiple views of the luggage to be obtained from an arrangement of linear X-ray detector arrays. This research investigated a solution to the unique problems defined when considering the possibility of replacing some of the X-ray sensor views with synthetic images. If sufficiently high quality synthetic images can be generated then intermediary X-ray sensors can be removed to minimise the hardware requirements and improve the commercial viability of the KDEX technique. Existing image synthesis algorithms are developed for visible light images. Due to fundamental differences between visible light and X-ray images, those algorithms are not directly applicable to the X-ray scenario. The conditions imposed by the X-ray images have instigated the original research and novel algorithm development and experimentation that form the body of this work. A voting based dual criteria multiple X-ray images synthesis algorithm (V-DMX) is proposed to exploit the potential of two matching criteria and information contained in a sequence of images. The V-DMX algorithm is divided into four stages
Recommended from our members
View synthesis for kinetic depth X-ray imaging
This thesis reports the development and analysis of feature based synthesis of transmission X-ray images. The synthetic imagery is formed through matching and morphing or warping line-scan format images produced by a novel multi-view X-ray machine. In this way video type sequences, which periodically alternate between synthetic and detector based views, may be formed. The purpose of these sequences is to provide depth from motion or kinetic depth effect (KDE) in a visual display; while the role of the synthesis is to reduce the total number of detector arrays, associated collimators and X-ray flux per inspection. A specific challenge is to explore the bounds for producing synthetic imagery that can be seamlessly introduced into the resultant sequences. This work is distinct from the image collection and display technique, termed KDEX, previously undertaken by the Imaging Science Group at NTU. The ultimate aim of the research programme in collaboration with The UK Home Office and The US Dept. of Homeland Security is to enhance the detection and identification of threats in X-ray scans of luggage. A multi-view „KDEX scanner‟ was employed to collect greyscale and colour coded image sequences of 30 different bags; each sequence comprised of 7 perspective views separated from one another by 10. This imagery was organised and stored in a database to enable a coherent series of experiments to be conducted. Corresponding features in sequential pairs of images, at various different angular separations, were identified by applying a scale invariant feature transform (SIFT)