4 research outputs found

    Scene and motion reconstruction from defocused and motion-blurred images via anisotropic diffusion

    No full text
    Abstract. We propose a solution to the problem of inferring the depth map, radiance and motion of a scene from a collection of motion-blurred and defocused images. We model motion-blur and defocus as an anisotropic diffusion process, whose initial conditions depend on the radiance and whose diffusion tensor encodes the shape of the scene, the motion field and the optics parameters. We show that this model is well-posed and propose an efficient algorithm to infer the unknowns of the model. Inference is performed by minimizing the discrepancy between the measured blurred images and the ones synthesized via forward diffusion. Since the problem is ill-posed, we also introduce additional Tikhonov regularization terms. The resulting method is fast and robust to noise as shown by experiments with both synthetic and real data.

    ์›€์ง์ด๋Š” ๋‹จ์ผ ์นด๋ฉ”๋ผ๋ฅผ ์ด์šฉํ•œ 3์ฐจ์› ๋ณต์›๊ณผ ๋””๋ธ”๋Ÿฌ๋ง, ์ดˆํ•ด์ƒ๋„ ๋ณต์›์˜ ๋™์‹œ์  ์ˆ˜ํ–‰ ๊ธฐ๋ฒ•

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2013. 8. ์ด๊ฒฝ๋ฌด.์˜์ƒ ๊ธฐ๋ฐ˜ 3์ฐจ์› ๋ณต์›์€ ์ปดํ“จํ„ฐ ๋น„์ „์˜ ๊ธฐ๋ณธ์ ์ธ ์—ฐ๊ตฌ ์ฃผ์ œ ๊ฐ€์šด๋ฐ ํ•˜๋‚˜๋กœ ์ตœ๊ทผ ๋ช‡ ๋…„๊ฐ„ ๋งŽ์€ ๋ฐœ์ „์ด ์žˆ์–ด์™”๋‹ค. ํŠนํžˆ ์ž๋™ ๋กœ๋ด‡์„ ์œ„ํ•œ ๋„ค๋น„๊ฒŒ์ด์…˜ ๋ฐ ํœด๋Œ€ ๊ธฐ๊ธฐ๋ฅผ ์ด์šฉํ•œ ์ฆ๊ฐ• ํ˜„์‹ค ๋“ฑ์— ๋„๋ฆฌ ํ™œ์šฉ๋  ์ˆ˜ ์žˆ๋Š” ๋‹จ์ผ ์นด๋ฉ”๋ผ๋ฅผ ์ด์šฉํ•œ 3์ฐจ์› ๋ณต์› ๊ธฐ๋ฒ•์€ ๋ณต์›์˜ ์ •ํ™•๋„, ๋ณต์› ๊ฐ€๋Šฅ ๋ฒ”์œ„ ๋ฐ ์ฒ˜๋ฆฌ ์†๋„ ์ธก๋ฉด์—์„œ ๋งŽ์€ ์‹ค์šฉ ๊ฐ€๋Šฅ์„ฑ์„ ๋ณด์—ฌ์ฃผ๊ณ  ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๊ทธ ์„ฑ๋Šฅ์€ ์—ฌ์ „ํžˆ ์กฐ์‹ฌ์Šค๋ ˆ ์ดฌ์˜๋œ ๋†’์€ ํ’ˆ์งˆ์˜ ์ž…๋ ฅ ์˜์ƒ์— ๋Œ€ํ•ด์„œ๋งŒ ์‹œํ—˜๋˜๊ณ  ์žˆ๋‹ค. ์›€์ง์ด๋Š” ๋‹จ์ผ ์นด๋ฉ”๋ผ๋ฅผ ์ด์šฉํ•œ 3์ฐจ์› ๋ณต์›์˜ ์‹ค์ œ ๋™์ž‘ ํ™˜๊ฒฝ์—์„œ๋Š” ์ž…๋ ฅ ์˜์ƒ์ด ํ™”์†Œ ์žก์Œ์ด๋‚˜ ์›€์ง์ž„์— ์˜ํ•œ ๋ฒˆ์ง ๋“ฑ์— ์˜ํ•˜์—ฌ ์†์ƒ๋  ์ˆ˜ ์žˆ๊ณ , ์˜์ƒ์˜ ํ•ด์ƒ๋„ ๋˜ํ•œ ์ •ํ™•ํ•œ ์นด๋ฉ”๋ผ ์œ„์น˜ ์ธ์‹ ๋ฐ 3์ฐจ์› ๋ณต์›์„ ์œ„ํ•ด์„œ๋Š” ์ถฉ๋ถ„ํžˆ ๋†’์ง€ ์•Š์„ ์ˆ˜ ์žˆ๋‹ค. ๋งŽ์€ ์—ฐ๊ตฌ์—์„œ ๊ณ ์„ฑ๋Šฅ ์˜์ƒ ํ™”์งˆ ํ–ฅ์ƒ ๊ธฐ๋ฒ•๋“ค์ด ์ œ์•ˆ๋˜์–ด ์™”์ง€๋งŒ ์ด๋“ค์€ ์ผ๋ฐ˜์ ์œผ๋กœ ๋†’์€ ๊ณ„์‚ฐ ๋น„์šฉ์„ ํ•„์š”๋กœ ํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์‹ค์‹œ๊ฐ„ ๋™์ž‘ ๋Šฅ๋ ฅ์ด ์ค‘์š”ํ•œ ๋‹จ์ผ ์นด๋ฉ”๋ผ ๊ธฐ๋ฐ˜ 3์ฐจ์› ๋ณต์›์— ์‚ฌ์šฉ๋˜๊ธฐ์—๋Š” ๋ถ€์ ํ•ฉํ•˜๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ๋ณด๋‹ค ์ •ํ™•ํ•˜๊ณ  ์•ˆ์ •๋œ ๋ณต์›์„ ์œ„ํ•˜์—ฌ ์˜์ƒ ๊ฐœ์„ ์ด ๊ฒฐํ•ฉ๋œ ์ƒˆ๋กœ์šด ๋‹จ์ผ ์นด๋ฉ”๋ผ ๊ธฐ๋ฐ˜ 3์ฐจ์› ๋ณต์› ๊ธฐ๋ฒ•์„ ๋‹ค๋ฃฌ๋‹ค. ์ด๋ฅผ ์œ„ํ•˜์—ฌ ์˜์ƒ ํ’ˆ์งˆ์ด ์ €ํ•˜๋˜๋Š” ์ค‘์š”ํ•œ ๋‘ ์š”์ธ์ธ ์›€์ง์ž„์— ์˜ํ•œ ์˜์ƒ ๋ฒˆ์ง๊ณผ ๋‚ฎ์€ ํ•ด์ƒ๋„ ๋ฌธ์ œ๊ฐ€ ๊ฐ๊ฐ ์  ๊ธฐ๋ฐ˜ ๋ณต์› ๋ฐ ์กฐ๋ฐ€ ๋ณต์› ๊ธฐ๋ฒ•๋“ค๊ณผ ๊ฒฐํ•ฉ๋œ๋‹ค. ์˜์ƒ ํ’ˆ์งˆ ์ €ํ•˜๋ฅผ ํฌํ•จํ•œ ์˜์ƒ ํš๋“ ๊ณผ์ •์€ ์นด๋ฉ”๋ผ ๋ฐ ์žฅ๋ฉด์˜ 3์ฐจ์› ๊ธฐํ•˜ ๊ตฌ์กฐ์™€ ๊ด€์ธก๋œ ์˜์ƒ ์‚ฌ์ด์˜ ๊ด€๊ณ„๋ฅผ ์ด์šฉํ•˜์—ฌ ๋ชจ๋ธ๋ง ํ•  ์ˆ˜ ์žˆ๊ณ , ์ด๋Ÿฌํ•œ ์˜์ƒ ํ’ˆ์งˆ ์ €ํ•˜ ๊ณผ์ •์„ ๊ณ ๋ คํ•จ์œผ๋กœ์จ ์ •ํ™•ํ•œ 3์ฐจ์› ๋ณต์›์„ ํ•˜๋Š” ๊ฒƒ์ด ๊ฐ€๋Šฅํ•ด์ง„๋‹ค. ๋˜ํ•œ, ์˜์ƒ ๋ฒˆ์ง ์ œ๊ฑฐ๋ฅผ ์œ„ํ•œ ๋ฒˆ์ง ์ปค๋„ ๋˜๋Š” ์˜์ƒ์˜ ์ดˆํ•ด์ƒ๋„ ๋ณต์›์„ ์œ„ํ•œ ํ™”์†Œ ๋Œ€์‘ ์ •๋ณด ๋“ฑ์ด 3์ฐจ์› ๋ณต์› ๊ณผ์ •๊ณผ ๋™์‹œ์— ์–ป์–ด์ง€๋Š”๊ฒƒ์ด ๊ฐ€๋Šฅํ•˜์—ฌ, ์˜์ƒ ๊ฐœ์„ ์ด ๋ณด๋‹ค ๊ฐ„ํŽธํ•˜๊ณ  ๋น ๋ฅด๊ฒŒ ์ˆ˜ํ–‰๋  ์ˆ˜ ์žˆ๋‹ค. ์ œ์•ˆ๋˜๋Š” ๊ธฐ๋ฒ•์€ 3์ฐจ์› ๋ณต์›๊ณผ ์˜์ƒ ๊ฐœ์„  ๋ฌธ์ œ๋ฅผ ๋™์‹œ์— ํ•ด๊ฒฐํ•จ์œผ๋กœ์จ ๊ฐ๊ฐ์˜ ๊ฒฐ๊ณผ๊ฐ€ ์ƒํ˜ธ ๋ณด์™„์ ์œผ๋กœ ํ–ฅ์ƒ๋œ๋‹ค๋Š” ์ ์—์„œ ๊ทธ ์žฅ์ ์„ ๊ฐ€์ง€๊ณ  ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์‹คํ—˜์  ํ‰๊ฐ€๋ฅผ ํ†ตํ•˜์—ฌ ์ œ์•ˆ๋˜๋Š” 3์ฐจ์› ๋ณต์› ๋ฐ ์˜์ƒ ๊ฐœ์„ ์˜ ํšจ๊ณผ์„ฑ์„ ์ž…์ฆํ•˜๋„๋ก ํ•œ๋‹ค.Vision-based 3D reconstruction is one of the fundamental problems in computer vision, and it has been researched intensively significantly in the last decades. In particular, 3D reconstruction using a single camera, which has a wide range of applications such as autonomous robot navigation and augmented reality, shows great possibilities in its reconstruction accuracy, scale of reconstruction coverage, and computational efficiency. However, until recently, the performances of most algorithms have been tested only with carefully recorded, high quality input sequences. In practical situations, input images for 3D reconstruction can be severely distorted due to various factors such as pixel noise and motion blur, and the resolution of images may not be high enough to achieve accurate camera localization and scene reconstruction results. Although various high-performance image enhancement methods have been proposed in many studies, the high computational costs of those methods prevent applying them to the 3D reconstruction systems where the real-time capability is an important issue. In this dissertation, novel single camera-based 3D reconstruction methods that are combined with image enhancement methods is studied to improve the accuracy and reliability of 3D reconstruction. To this end, two critical image degradations, motion blur and low image resolution, are addressed for both sparse reconstruction and dense 3D reconstruction systems, and novel integrated enhancement methods for those degradations are presented. Using the relationship between the observed images and 3D geometry of the camera and scenes, the image formation process including image degradations is modeled by the camera and scene geometry. Then, by taking the image degradation factors in consideration, accurate 3D reconstruction then is achieved. Furthermore, the information required for image enhancement, such as blur kernels for deblurring and pixel correspondences for super-resolution, is simultaneously obtained while reconstructing 3D scene, and this makes the image enhancement much simpler and faster. The proposed methods have an advantage that the results of 3D reconstruction and image enhancement are improved by each other with the simultaneous solution of these problems. Experimental evaluations demonstrate the effectiveness of the proposed 3D reconstruction and image enhancement methods.1. Introduction 2. Sparse 3D Reconstruction and Image Deblurring 3. Sparse 3D Reconstruction and Image Super-Resolution 4. Dense 3D Reconstruction and Image Deblurring 5. Dense 3D Reconstruction and Image Super-Resolution 6. Dense 3D Reconstruction, Image Deblurring, and Super-Resolution 7. ConclusionDocto

    Joint methods in imaging based on diffuse image representations

    Get PDF
    This thesis deals with the application and the analysis of different variants of the Mumford-Shah model in the context of image processing. In this kind of models, a given function is approximated in a piecewise smooth or piecewise constant manner. Especially the numerical treatment of the discontinuities requires additional models that are also outlined in this work. The main part of this thesis is concerned with four different topics. Simultaneous edge detection and registration of two images: The image edges are detected with the Ambrosio-Tortorelli model, an approximation of the Mumford-Shah model that approximates the discontinuity set with a phase field, and the registration is based on these edges. The registration obtained by this model is fully symmetric in the sense that the same matching is obtained if the roles of the two input images are swapped. Detection of grain boundaries from atomic scale images of metals or metal alloys: This is an image processing problem from materials science where atomic scale images are obtained either experimentally for instance by transmission electron microscopy or by numerical simulation tools. Grains are homogenous material regions whose atomic lattice orientation differs from their surroundings. Based on a Mumford-Shah type functional, the grain boundaries are modeled as the discontinuity set of the lattice orientation. In addition to the grain boundaries, the model incorporates the extraction of a global elastic deformation of the atomic lattice. Numerically, the discontinuity set is modeled by a level set function following the approach by Chan and Vese. Joint motion estimation and restoration of motion-blurred video: A variational model for joint object detection, motion estimation and deblurring of consecutive video frames is proposed. For this purpose, a new motion blur model is developed that accurately describes the blur also close to the boundary of a moving object. Here, the video is assumed to consist of an object moving in front of a static background. The segmentation into object and background is handled by a Mumford-Shah type aspect of the proposed model. Convexification of the binary Mumford-Shah segmentation model: After considering the application of Mumford-Shah type models to tackle specific image processing problems in the previous topics, the Mumford-Shah model itself is studied more closely. Inspired by the work of Nikolova, Esedoglu and Chan, a method is developed that allows global minimization of the binary Mumford-Shah segmentation model by solving a convex, unconstrained optimization problem. In an outlook, segmentation of flowfields into piecewise affine regions using this convexification method is briefly discussed

    Modeling and applications of the focus cue in conventional digital cameras

    Get PDF
    El enfoque en cรกmaras digitales juega un papel fundamental tanto en la calidad de la imagen como en la percepciรณn del entorno. Esta tesis estudia el enfoque en cรกmaras digitales convencionales, tales como cรกmaras de mรณviles, fotogrรกficas, webcams y similares. Una revisiรณn rigurosa de los conceptos teรณricos detras del enfoque en cรกmaras convencionales muestra que, a pasar de su utilidad, el modelo clรกsico del thin lens presenta muchas limitaciones para aplicaciรณn en diferentes problemas relacionados con el foco. En esta tesis, el focus profile es propuesto como una alternativa a conceptos clรกsicos como la profundidad de campo. Los nuevos conceptos introducidos en esta tesis son aplicados a diferentes problemas relacionados con el foco, tales como la adquisiciรณn eficiente de imรกgenes, estimaciรณn de profundidad, integraciรณn de elementos perceptuales y fusiรณn de imรกgenes. Los resultados experimentales muestran la aplicaciรณn exitosa de los modelos propuestos.The focus of digital cameras plays a fundamental role in both the quality of the acquired images and the perception of the imaged scene. This thesis studies the focus cue in conventional cameras with focus control, such as cellphone cameras, photography cameras, webcams and the like. A deep review of the theoretical concepts behind focus in conventional cameras reveals that, despite its usefulness, the widely known thin lens model has several limitations for solving different focus-related problems in computer vision. In order to overcome these limitations, the focus profile model is introduced as an alternative to classic concepts, such as the near and far limits of the depth-of-field. The new concepts introduced in this dissertation are exploited for solving diverse focus-related problems, such as efficient image capture, depth estimation, visual cue integration and image fusion. The results obtained through an exhaustive experimental validation demonstrate the applicability of the proposed models
    corecore