13,757 research outputs found
Recent Progress in Image Deblurring
This paper comprehensively reviews the recent development of image
deblurring, including non-blind/blind, spatially invariant/variant deblurring
techniques. Indeed, these techniques share the same objective of inferring a
latent sharp image from one or several corresponding blurry images, while the
blind deblurring techniques are also required to derive an accurate blur
kernel. Considering the critical role of image restoration in modern imaging
systems to provide high-quality images under complex environments such as
motion, undesirable lighting conditions, and imperfect system components, image
deblurring has attracted growing attention in recent years. From the viewpoint
of how to handle the ill-posedness which is a crucial issue in deblurring
tasks, existing methods can be grouped into five categories: Bayesian inference
framework, variational methods, sparse representation-based methods,
homography-based modeling, and region-based methods. In spite of achieving a
certain level of development, image deblurring, especially the blind case, is
limited in its success by complex application conditions which make the blur
kernel hard to obtain and be spatially variant. We provide a holistic
understanding and deep insight into image deblurring in this review. An
analysis of the empirical evidence for representative methods, practical
issues, as well as a discussion of promising future directions are also
presented.Comment: 53 pages, 17 figure
Learning Optimization-inspired Image Propagation with Control Mechanisms and Architecture Augmentations for Low-level Vision
In recent years, building deep learning models from optimization perspectives
has becoming a promising direction for solving low-level vision problems. The
main idea of most existing approaches is to straightforwardly combine numerical
iterations with manually designed network architectures to generate image
propagations for specific kinds of optimization models. However, these
heuristic learning models often lack mechanisms to control the propagation and
rely on architecture engineering heavily. To mitigate the above issues, this
paper proposes a unified optimization-inspired deep image propagation framework
to aggregate Generative, Discriminative and Corrective (GDC for short)
principles for a variety of low-level vision tasks. Specifically, we first
formulate low-level vision tasks using a generic optimization objective and
construct our fundamental propagative modules from three different viewpoints,
i.e., the solution could be obtained/learned 1) in generative manner; 2) based
on discriminative metric, and 3) with domain knowledge correction. By designing
control mechanisms to guide image propagations, we then obtain convergence
guarantees of GDC for both fully- and partially-defined optimization
formulations. Furthermore, we introduce two architecture augmentation
strategies (i.e., normalization and automatic search) to respectively enhance
the propagation stability and task/data-adaption ability. Extensive experiments
on different low-level vision applications demonstrate the effectiveness and
flexibility of GDC.Comment: 15 page
Image enhancement methods and applications in computational photography
Computational photography is currently a rapidly developing and cutting-edge topic in applied optics, image sensors and image processing fields to go beyond the limitations of traditional photography. The innovations of computational photography allow the photographer not only merely to take an image, but also, more importantly, to perform computations on the captured image data. Good examples of these innovations include high dynamic range imaging, focus stacking, super-resolution, motion deblurring and so on. Although extensive work has been done to explore image enhancement techniques in each subfield of computational photography, attention has seldom been given to study of the image enhancement technique of simultaneously extending depth of field and dynamic range of a scene. In my dissertation, I present an algorithm which combines focus stacking and high dynamic range (HDR) imaging in order to produce an image with both extended depth of field (DOF) and dynamic range than any of the input images. In this dissertation, I also investigate super-resolution image restoration from multiple images, which are possibly degraded by large motion blur. The proposed algorithm combines the super-resolution problem and blind image deblurring problem in a unified framework. The blur kernel for each input image is separately estimated. I also do not make any restrictions on the motion fields among images; that is, I estimate dense motion field without simplifications such as parametric motion. While the proposed super-resolution method uses multiple images to enhance spatial resolution from multiple regular images, single image super-resolution is related to techniques of denoising or removing blur from one single captured image. In my dissertation, space-varying point spread function (PSF) estimation and image deblurring for single image is also investigated. Regarding the PSF estimation, I do not make any restrictions on the type of blur or how the blur varies spatially. Once the space-varying PSF is estimated, space-varying image deblurring is performed, which produces good results even for regions where it is not clear what the correct PSF is at first. I also bring image enhancement applications to both personal computer (PC) and Android platform as computational photography applications
A Theoretically Guaranteed Deep Optimization Framework for Robust Compressive Sensing MRI
Magnetic Resonance Imaging (MRI) is one of the most dynamic and safe imaging
techniques available for clinical applications. However, the rather slow speed
of MRI acquisitions limits the patient throughput and potential indi cations.
Compressive Sensing (CS) has proven to be an efficient technique for
accelerating MRI acquisition. The most widely used CS-MRI model, founded on the
premise of reconstructing an image from an incompletely filled k-space, leads
to an ill-posed inverse problem. In the past years, lots of efforts have been
made to efficiently optimize the CS-MRI model. Inspired by deep learning
techniques, some preliminary works have tried to incorporate deep architectures
into CS-MRI process. Unfortunately, the convergence issues (due to the
experience-based networks) and the robustness (i.e., lack real-world noise
modeling) of these deeply trained optimization methods are still missing. In
this work, we develop a new paradigm to integrate designed numerical solvers
and the data-driven architectures for CS-MRI. By introducing an optimal
condition checking mechanism, we can successfully prove the convergence of our
established deep CS-MRI optimization scheme. Furthermore, we explicitly
formulate the Rician noise distributions within our framework and obtain an
extended CS-MRI network to handle the real-world nosies in the MRI process.
Extensive experimental results verify that the proposed paradigm outperforms
the existing state-of-the-art techniques both in reconstruction accuracy and
efficiency as well as robustness to noises in real scene
Fast restoration for out-of-focus blurred images of QR code with edge prior information via image sensing.
Out-of-focus blurring of the QR code is very common in mobile Internet systems, which often causes failure of authentication as a result of a misreading of the information hence adversely affects the operation of the system. To tackle this difficulty, this work firstly introduced an edge prior information, which is the average distance between the center point and the edge of the clear QR code images in the same batch. It is motivated by the theoretical analysis and the practical observation of the theory of CMOS image sensing, optics information, blur invariants, and the invariance of the center of the diffuse light spots. After obtaining the edge prior information, combining the iterative image and the center point of the binary image, the proposed method can accurately estimate the parameter of the out-of-focus blur kernel. Furthermore, we obtain the sharp image by Wiener filter, a non-blind image deblurring algorithm. By this, it avoids excessive redundant calculations. Experimental results validate that the proposed method has great practical utility in terms of deblurring quality, robustness, and computational efficiency, which is suitable for barcode application systems, e.g., warehouse, logistics, and automated production
Image Deblurring and Near-real-time Atmospheric Seeing Estimation through the Employment of Convergence of Variance
A new image reconstruction algorithm is presented that will remove the effect of atmospheric turbulence on motion compensated frame average images. The primary focus of this research was to develop a blind deconvolution technique that could be employed in a tactical military environment where both time and computational power are limited. Additionally, this technique can be employed to measure atmospheric seeing conditions. In a blind deconvolution fashion, the algorithm simultaneously computes a high resolution image and an average model for the atmospheric blur parameterized by Fried’s seeing parameter. The difference in this approach is that it does not assume a prior distribution for the seeing parameter, rather it assesses the convergence of the image’s variance as the stopping criteria and identification of the proper seeing parameter from a range of candidate values. Experimental results show that the convergence of variance technique allows for estimation of the seeing parameter accurate to within 0.5 cm and often even better depending on the signal to noise ratio
Motion Offset for Blur Modeling
Motion blur caused by the relative movement between the camera and the subject is often an undesirable degradation of the image quality. In most conventional deblurring methods, a blur kernel is estimated for image deconvolution. Due to the ill-posed nature, predefined priors are proposed to suppress the ill-posedness. However, these predefined priors can only handle some specific situations. In order to achieve a better deblurring performance on dynamic scene, deep-learning based methods are proposed to learn a mapping function that restore the sharp image from a blurry image. The blur may be implicitly modelled in feature extraction module. However, the blur modelled from the paired dataset cannot be well generalized to some real-world scenes. To summary, an accurate and dynamic blur model that more closely approximates real-world blur is needed.
By revisiting the principle of camera exposure, we can model the blur with the displacements between sharp pixels and the exposed pixel, namely motion offsets. Given specific physical constraints, motion offsets are able to form different exposure trajectories (i.e. linear, quadratic). Compare to conventional blur kernel, our proposed motion offsets are a more rigorous approximation for real-world blur, since they can constitute a non-linear and non-uniform motion field. Through learning from dynamic scene dataset, an accurate and spatial-variant motion offset field is obtained.
With accurate motion information and a compact blur modeling method, we explore the ways of utilizing motion information to facilitate multiple blur-related tasks. By introducing recovered motion offsets, we build up a motion-aware and spatial-variant convolution. For extracting a video clip from a blurry image, motion offsets can provide an explicit (non-)linear motion trajectory for interpolating. We also work towards a better image deblurring performance in real-world scenarios by improving the generalization ability of the deblurring model
- …