6 research outputs found

    ํ†ตํ•ฉ์‹œ์Šคํ…œ์„์ด์šฉํ•œ ๋‹ค์‹œ์ ์Šคํ…Œ๋ ˆ์˜ค ๋งค์นญ๊ณผ์˜์ƒ๋ณต์›

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2017. 2. ์ด๊ฒฝ๋ฌด.Estimating camera pose and scene structures from seriously degraded images is challenging problem. Most existing multi-view stereo algorithms assume high-quality input images and therefore have unreliable results for blurred, noisy, or low-resolution images. Experimental results show that the approach of using off-the-shelf image reconstruction algorithms as independent preprocessing is generally ineffective or even sometimes counterproductive. This is because naive frame-wise image reconstruction methods fundamentally ignore the consistency between images, although they seem to produce visually plausible results. In this thesis, from the fact that image reconstruction and multi-view stereo problems are interrelated, we present a unified framework to solve these problems jointly. The validity of this approach is empirically verified for four different problems, dense depth map reconstruction, camera pose estimation, super-resolution, and deblurring from images obtained by a single moving camera. By reflecting the physical imaging process, we cast our objective into a cost minimization problem, and solve the solution using alternate optimization techniques. Experiments show that the proposed method can restore high-quality depth maps from seriously degraded images for both synthetic and real video, as opposed to the failure of simple multi-view stereo methods. Our algorithm also produces superior super-resolution and deblurring results compared to simple preprocessing with conventional super-resolution and deblurring techniques. Moreover, we show that the proposed framework can be generalized to handle more common scenarios. First, it can solve image reconstruction and multi-view stereo problems for multi-view single-shot images captured by a light field camera. By using information of calibrated multi-view images, it recovers the motions of individual objects in the input image as well as the unknown camera motion during the shutter time. The contribution of this thesis is proposing a new perspective on the solution of the existing computer vision problems from an integrated viewpoint. We show that by solving interrelated problems jointly, we can obtain physically more plausible solution and better performance, especially when input images are challenging. The proposed optimization algorithm also makes our algorithm more practical in terms of computational complexity.1 Introduction 1 1.1 Outline of Dissertation 2 2 Background 5 3 Generalized Imaging Model 9 3.1 Camera Projection Model 9 3.2 Depth and Warping Operation 11 3.3 Representation of Camera Pose in SE(3) 12 3.4 Proposed Imaging Model 12 4 Rendering Synthetic Datasets 17 4.1 Making Blurred Image Sequences using Depth-based Image Rendering 18 4.2 Making Blurred Image Sequences using Blender 18 5 A Unified Framework for Single-shot Multi-view Images 21 5.1 Introduction 21 5.2 Related Works 24 5.3 Deblurring with 4D Light Fields 27 5.3.1 Motion Blur Formulation in Light Fields 27 5.3.2 Initialization 28 5.4 Joint Estimation 30 5.4.1 Energy Formulation 30 5.4.2 Update Latent Image 31 5.4.3 Update Camera Pose and Depth map 33 5.5 Experimental Results 34 5.5.1 Synthetic Data 34 5.5.2 Real Data 36 5.6 Conclusion 37 6 A Unified Framework for a Monocular Image Sequence 41 6.1 Introduction 41 6.2 Related Works 44 6.3 Modeling Imaging Process 46 6.4 Unified Energy Formulation 47 6.4.1 Matching term 47 6.4.2 Self-consistency term 48 6.4.3 Regularization term 49 6.5 Optimization 50 6.5.1 Update of the depth maps and camera poses 51 6.5.2 Update of the latent images . 52 6.5.3 Initialization 53 6.5.4 Occlusion Handling 54 6.6 Experimental Results 54 6.6.1 Synthetic datasets 55 6.6.2 Real datasets 61 6.6.3 The effect of parameters 65 6.7 Conclusion 66 7 A Unified Framework for SLAM 69 7.1 Motivation 69 7.2 Baseline 70 7.3 Proposed Method 72 7.4 Experimental Results 73 7.4.1 Quantitative comparison 73 7.4.2 Qualitative results 77 7.4.3 Runtime 79 7.5 Conclusion 80 8 Conclusion 83 8.1 Summary and Contribution of the Dissertation 83 8.2 Future Works 84 Bibliography 86 ์ดˆ๋ก 94Docto

    ์›€์ง์ด๋Š” ๋‹จ์ผ ์นด๋ฉ”๋ผ๋ฅผ ์ด์šฉํ•œ 3์ฐจ์› ๋ณต์›๊ณผ ๋””๋ธ”๋Ÿฌ๋ง, ์ดˆํ•ด์ƒ๋„ ๋ณต์›์˜ ๋™์‹œ์  ์ˆ˜ํ–‰ ๊ธฐ๋ฒ•

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2013. 8. ์ด๊ฒฝ๋ฌด.์˜์ƒ ๊ธฐ๋ฐ˜ 3์ฐจ์› ๋ณต์›์€ ์ปดํ“จํ„ฐ ๋น„์ „์˜ ๊ธฐ๋ณธ์ ์ธ ์—ฐ๊ตฌ ์ฃผ์ œ ๊ฐ€์šด๋ฐ ํ•˜๋‚˜๋กœ ์ตœ๊ทผ ๋ช‡ ๋…„๊ฐ„ ๋งŽ์€ ๋ฐœ์ „์ด ์žˆ์–ด์™”๋‹ค. ํŠนํžˆ ์ž๋™ ๋กœ๋ด‡์„ ์œ„ํ•œ ๋„ค๋น„๊ฒŒ์ด์…˜ ๋ฐ ํœด๋Œ€ ๊ธฐ๊ธฐ๋ฅผ ์ด์šฉํ•œ ์ฆ๊ฐ• ํ˜„์‹ค ๋“ฑ์— ๋„๋ฆฌ ํ™œ์šฉ๋  ์ˆ˜ ์žˆ๋Š” ๋‹จ์ผ ์นด๋ฉ”๋ผ๋ฅผ ์ด์šฉํ•œ 3์ฐจ์› ๋ณต์› ๊ธฐ๋ฒ•์€ ๋ณต์›์˜ ์ •ํ™•๋„, ๋ณต์› ๊ฐ€๋Šฅ ๋ฒ”์œ„ ๋ฐ ์ฒ˜๋ฆฌ ์†๋„ ์ธก๋ฉด์—์„œ ๋งŽ์€ ์‹ค์šฉ ๊ฐ€๋Šฅ์„ฑ์„ ๋ณด์—ฌ์ฃผ๊ณ  ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๊ทธ ์„ฑ๋Šฅ์€ ์—ฌ์ „ํžˆ ์กฐ์‹ฌ์Šค๋ ˆ ์ดฌ์˜๋œ ๋†’์€ ํ’ˆ์งˆ์˜ ์ž…๋ ฅ ์˜์ƒ์— ๋Œ€ํ•ด์„œ๋งŒ ์‹œํ—˜๋˜๊ณ  ์žˆ๋‹ค. ์›€์ง์ด๋Š” ๋‹จ์ผ ์นด๋ฉ”๋ผ๋ฅผ ์ด์šฉํ•œ 3์ฐจ์› ๋ณต์›์˜ ์‹ค์ œ ๋™์ž‘ ํ™˜๊ฒฝ์—์„œ๋Š” ์ž…๋ ฅ ์˜์ƒ์ด ํ™”์†Œ ์žก์Œ์ด๋‚˜ ์›€์ง์ž„์— ์˜ํ•œ ๋ฒˆ์ง ๋“ฑ์— ์˜ํ•˜์—ฌ ์†์ƒ๋  ์ˆ˜ ์žˆ๊ณ , ์˜์ƒ์˜ ํ•ด์ƒ๋„ ๋˜ํ•œ ์ •ํ™•ํ•œ ์นด๋ฉ”๋ผ ์œ„์น˜ ์ธ์‹ ๋ฐ 3์ฐจ์› ๋ณต์›์„ ์œ„ํ•ด์„œ๋Š” ์ถฉ๋ถ„ํžˆ ๋†’์ง€ ์•Š์„ ์ˆ˜ ์žˆ๋‹ค. ๋งŽ์€ ์—ฐ๊ตฌ์—์„œ ๊ณ ์„ฑ๋Šฅ ์˜์ƒ ํ™”์งˆ ํ–ฅ์ƒ ๊ธฐ๋ฒ•๋“ค์ด ์ œ์•ˆ๋˜์–ด ์™”์ง€๋งŒ ์ด๋“ค์€ ์ผ๋ฐ˜์ ์œผ๋กœ ๋†’์€ ๊ณ„์‚ฐ ๋น„์šฉ์„ ํ•„์š”๋กœ ํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์‹ค์‹œ๊ฐ„ ๋™์ž‘ ๋Šฅ๋ ฅ์ด ์ค‘์š”ํ•œ ๋‹จ์ผ ์นด๋ฉ”๋ผ ๊ธฐ๋ฐ˜ 3์ฐจ์› ๋ณต์›์— ์‚ฌ์šฉ๋˜๊ธฐ์—๋Š” ๋ถ€์ ํ•ฉํ•˜๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ๋ณด๋‹ค ์ •ํ™•ํ•˜๊ณ  ์•ˆ์ •๋œ ๋ณต์›์„ ์œ„ํ•˜์—ฌ ์˜์ƒ ๊ฐœ์„ ์ด ๊ฒฐํ•ฉ๋œ ์ƒˆ๋กœ์šด ๋‹จ์ผ ์นด๋ฉ”๋ผ ๊ธฐ๋ฐ˜ 3์ฐจ์› ๋ณต์› ๊ธฐ๋ฒ•์„ ๋‹ค๋ฃฌ๋‹ค. ์ด๋ฅผ ์œ„ํ•˜์—ฌ ์˜์ƒ ํ’ˆ์งˆ์ด ์ €ํ•˜๋˜๋Š” ์ค‘์š”ํ•œ ๋‘ ์š”์ธ์ธ ์›€์ง์ž„์— ์˜ํ•œ ์˜์ƒ ๋ฒˆ์ง๊ณผ ๋‚ฎ์€ ํ•ด์ƒ๋„ ๋ฌธ์ œ๊ฐ€ ๊ฐ๊ฐ ์  ๊ธฐ๋ฐ˜ ๋ณต์› ๋ฐ ์กฐ๋ฐ€ ๋ณต์› ๊ธฐ๋ฒ•๋“ค๊ณผ ๊ฒฐํ•ฉ๋œ๋‹ค. ์˜์ƒ ํ’ˆ์งˆ ์ €ํ•˜๋ฅผ ํฌํ•จํ•œ ์˜์ƒ ํš๋“ ๊ณผ์ •์€ ์นด๋ฉ”๋ผ ๋ฐ ์žฅ๋ฉด์˜ 3์ฐจ์› ๊ธฐํ•˜ ๊ตฌ์กฐ์™€ ๊ด€์ธก๋œ ์˜์ƒ ์‚ฌ์ด์˜ ๊ด€๊ณ„๋ฅผ ์ด์šฉํ•˜์—ฌ ๋ชจ๋ธ๋ง ํ•  ์ˆ˜ ์žˆ๊ณ , ์ด๋Ÿฌํ•œ ์˜์ƒ ํ’ˆ์งˆ ์ €ํ•˜ ๊ณผ์ •์„ ๊ณ ๋ คํ•จ์œผ๋กœ์จ ์ •ํ™•ํ•œ 3์ฐจ์› ๋ณต์›์„ ํ•˜๋Š” ๊ฒƒ์ด ๊ฐ€๋Šฅํ•ด์ง„๋‹ค. ๋˜ํ•œ, ์˜์ƒ ๋ฒˆ์ง ์ œ๊ฑฐ๋ฅผ ์œ„ํ•œ ๋ฒˆ์ง ์ปค๋„ ๋˜๋Š” ์˜์ƒ์˜ ์ดˆํ•ด์ƒ๋„ ๋ณต์›์„ ์œ„ํ•œ ํ™”์†Œ ๋Œ€์‘ ์ •๋ณด ๋“ฑ์ด 3์ฐจ์› ๋ณต์› ๊ณผ์ •๊ณผ ๋™์‹œ์— ์–ป์–ด์ง€๋Š”๊ฒƒ์ด ๊ฐ€๋Šฅํ•˜์—ฌ, ์˜์ƒ ๊ฐœ์„ ์ด ๋ณด๋‹ค ๊ฐ„ํŽธํ•˜๊ณ  ๋น ๋ฅด๊ฒŒ ์ˆ˜ํ–‰๋  ์ˆ˜ ์žˆ๋‹ค. ์ œ์•ˆ๋˜๋Š” ๊ธฐ๋ฒ•์€ 3์ฐจ์› ๋ณต์›๊ณผ ์˜์ƒ ๊ฐœ์„  ๋ฌธ์ œ๋ฅผ ๋™์‹œ์— ํ•ด๊ฒฐํ•จ์œผ๋กœ์จ ๊ฐ๊ฐ์˜ ๊ฒฐ๊ณผ๊ฐ€ ์ƒํ˜ธ ๋ณด์™„์ ์œผ๋กœ ํ–ฅ์ƒ๋œ๋‹ค๋Š” ์ ์—์„œ ๊ทธ ์žฅ์ ์„ ๊ฐ€์ง€๊ณ  ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์‹คํ—˜์  ํ‰๊ฐ€๋ฅผ ํ†ตํ•˜์—ฌ ์ œ์•ˆ๋˜๋Š” 3์ฐจ์› ๋ณต์› ๋ฐ ์˜์ƒ ๊ฐœ์„ ์˜ ํšจ๊ณผ์„ฑ์„ ์ž…์ฆํ•˜๋„๋ก ํ•œ๋‹ค.Vision-based 3D reconstruction is one of the fundamental problems in computer vision, and it has been researched intensively significantly in the last decades. In particular, 3D reconstruction using a single camera, which has a wide range of applications such as autonomous robot navigation and augmented reality, shows great possibilities in its reconstruction accuracy, scale of reconstruction coverage, and computational efficiency. However, until recently, the performances of most algorithms have been tested only with carefully recorded, high quality input sequences. In practical situations, input images for 3D reconstruction can be severely distorted due to various factors such as pixel noise and motion blur, and the resolution of images may not be high enough to achieve accurate camera localization and scene reconstruction results. Although various high-performance image enhancement methods have been proposed in many studies, the high computational costs of those methods prevent applying them to the 3D reconstruction systems where the real-time capability is an important issue. In this dissertation, novel single camera-based 3D reconstruction methods that are combined with image enhancement methods is studied to improve the accuracy and reliability of 3D reconstruction. To this end, two critical image degradations, motion blur and low image resolution, are addressed for both sparse reconstruction and dense 3D reconstruction systems, and novel integrated enhancement methods for those degradations are presented. Using the relationship between the observed images and 3D geometry of the camera and scenes, the image formation process including image degradations is modeled by the camera and scene geometry. Then, by taking the image degradation factors in consideration, accurate 3D reconstruction then is achieved. Furthermore, the information required for image enhancement, such as blur kernels for deblurring and pixel correspondences for super-resolution, is simultaneously obtained while reconstructing 3D scene, and this makes the image enhancement much simpler and faster. The proposed methods have an advantage that the results of 3D reconstruction and image enhancement are improved by each other with the simultaneous solution of these problems. Experimental evaluations demonstrate the effectiveness of the proposed 3D reconstruction and image enhancement methods.1. Introduction 2. Sparse 3D Reconstruction and Image Deblurring 3. Sparse 3D Reconstruction and Image Super-Resolution 4. Dense 3D Reconstruction and Image Deblurring 5. Dense 3D Reconstruction and Image Super-Resolution 6. Dense 3D Reconstruction, Image Deblurring, and Super-Resolution 7. ConclusionDocto

    Super-resolution of 3-dimensional scenes

    Full text link
    Super-resolution is an image enhancement method that increases the resolution of images and video. Previously this technique could only be applied to 2D scenes. The super-resolution algorithm developed in this thesis creates high-resolution views of 3-dimensional scenes, using low-resolution images captured from varying, unknown positions
    corecore