17 research outputs found
Learning to Estimate and Remove Non-uniform Image Blur
International audienceThis paper addresses the problem of restoring images subjected to unknown and spatially varying blur caused by defocus or linear (say, horizontal) motion. The estimation of the global (non-uniform) image blur is cast as a multi-label energy minimization problem. The energy is the sum of unary terms corresponding to learned local blur estimators, and binary ones corresponding to blur smoothness. Its global minimum is found using Ishikawa's method by exploiting the natural order of discretized blur values for linear motions and defocus. Once the blur has been estimated, the image is restored using a robust (non-uniform) deblurring algorithm based on sparse regularization with global image statistics. The proposed algorithm outputs both a segmentation of the image into uniform-blur layers and an estimate of the corresponding sharp image. We present qualitative results on real images, and use synthetic data to quantitatively compare our approach to the publicly available implementation of Chakrabarti et al. (2010)
Motion Deblurring in the Wild
The task of image deblurring is a very ill-posed problem as both the image
and the blur are unknown. Moreover, when pictures are taken in the wild, this
task becomes even more challenging due to the blur varying spatially and the
occlusions between the object. Due to the complexity of the general image model
we propose a novel convolutional network architecture which directly generates
the sharp image.This network is built in three stages, and exploits the
benefits of pyramid schemes often used in blind deconvolution. One of the main
difficulties in training such a network is to design a suitable dataset. While
useful data can be obtained by synthetically blurring a collection of images,
more realistic data must be collected in the wild. To obtain such data we use a
high frame rate video camera and keep one frame as the sharp image and frame
average as the corresponding blurred image. We show that this realistic dataset
is key in achieving state-of-the-art performance and dealing with occlusions
Modeling Camera Effects to Improve Visual Learning from Synthetic Data
Recent work has focused on generating synthetic imagery to increase the size
and variability of training data for learning visual tasks in urban scenes.
This includes increasing the occurrence of occlusions or varying environmental
and weather effects. However, few have addressed modeling variation in the
sensor domain. Sensor effects can degrade real images, limiting
generalizability of network performance on visual tasks trained on synthetic
data and tested in real environments. This paper proposes an efficient,
automatic, physically-based augmentation pipeline to vary sensor effects
--chromatic aberration, blur, exposure, noise, and color cast-- for synthetic
imagery. In particular, this paper illustrates that augmenting synthetic
training datasets with the proposed pipeline reduces the domain gap between
synthetic and real domains for the task of object detection in urban driving
scenes
Learning a Convolutional Neural Network for Non-uniform Motion Blur Removal
In this paper, we address the problem of estimating and removing non-uniform
motion blur from a single blurry image. We propose a deep learning approach to
predicting the probabilistic distribution of motion blur at the patch level
using a convolutional neural network (CNN). We further extend the candidate set
of motion kernels predicted by the CNN using carefully designed image
rotations. A Markov random field model is then used to infer a dense
non-uniform motion blur field enforcing motion smoothness. Finally, motion blur
is removed by a non-uniform deblurring model using patch-level image prior.
Experimental evaluations show that our approach can effectively estimate and
remove complex non-uniform motion blur that is not handled well by previous
approaches.Comment: This is a final version accepted by CVPR 201
End-to-end Interpretable Learning of Non-blind Image Deblurring
Non-blind image deblurring is typically formulated as a linear least-squares
problem regularized by natural priors on the corresponding sharp picture's
gradients, which can be solved, for example, using a half-quadratic splitting
method with Richardson fixed-point iterations for its least-squares updates and
a proximal operator for the auxiliary variable updates. We propose to
precondition the Richardson solver using approximate inverse filters of the
(known) blur and natural image prior kernels. Using convolutions instead of a
generic linear preconditioner allows extremely efficient parameter sharing
across the image, and leads to significant gains in accuracy and/or speed
compared to classical FFT and conjugate-gradient methods. More importantly, the
proposed architecture is easily adapted to learning both the preconditioner and
the proximal operator using CNN embeddings. This yields a simple and efficient
algorithm for non-blind image deblurring which is fully interpretable, can be
learned end to end, and whose accuracy matches or exceeds the state of the art,
quite significantly, in the non-uniform case.Comment: Accepted at ECCV2020 (poster