2,699 research outputs found
MSR-net:Low-light Image Enhancement Using Deep Convolutional Network
Images captured in low-light conditions usually suffer from very low
contrast, which increases the difficulty of subsequent computer vision tasks in
a great extent. In this paper, a low-light image enhancement model based on
convolutional neural network and Retinex theory is proposed. Firstly, we show
that multi-scale Retinex is equivalent to a feedforward convolutional neural
network with different Gaussian convolution kernels. Motivated by this fact, we
consider a Convolutional Neural Network(MSR-net) that directly learns an
end-to-end mapping between dark and bright images. Different fundamentally from
existing approaches, low-light image enhancement in this paper is regarded as a
machine learning problem. In this model, most of the parameters are optimized
by back-propagation, while the parameters of traditional models depend on the
artificial setting. Experiments on a number of challenging images reveal the
advantages of our method in comparison with other state-of-the-art methods from
the qualitative and quantitative perspective.Comment: 9page
Measure-Valued Variational Models with Applications to Diffusion-Weighted Imaging
We develop a general mathematical framework for variational problems where
the unknown function assumes values in the space of probability measures on
some metric space. We study weak and strong topologies and define a total
variation seminorm for functions taking values in a Banach space. The seminorm
penalizes jumps and is rotationally invariant under certain conditions. We
prove existence of a minimizer for a class of variational problems based on
this formulation of total variation, and provide an example where uniqueness
fails to hold. Employing the Kan\-torovich-Rubinstein transport norm from the
theory of optimal transport, we propose a variational approach for the
restoration of orientation distribution function (ODF)-valued images, as
commonly used in Diffusion MRI. We demonstrate that the approach is numerically
feasible on several data sets.Comment: Accepted by Journal of Mathematical Imaging and Vision (SSVM 2017
special issue
Fractional Multiscale Fusion-based De-hazing
This report presents the results of a proposed multi-scale fusion-based
single image de-hazing algorithm, which can also be used for underwater image
enhancement. Furthermore, the algorithm was designed for very fast operation
and minimal run-time. The proposed scheme is the faster than existing
algorithms for both de-hazing and underwater image enhancement and amenable to
digital hardware implementation. Results indicate mostly consistent and good
results for both categories of images when compared with other algorithms from
the literature.Comment: 23 pages, 13 figures, 2 table
Generation of High Dynamic Range Illumination from a Single Image for the Enhancement of Undesirably Illuminated Images
This paper presents an algorithm that enhances undesirably illuminated images
by generating and fusing multi-level illuminations from a single image.The
input image is first decomposed into illumination and reflectance components by
using an edge-preserving smoothing filter. Then the reflectance component is
scaled up to improve the image details in bright areas. The illumination
component is scaled up and down to generate several illumination images that
correspond to certain camera exposure values different from the original. The
virtual multi-exposure illuminations are blended into an enhanced illumination,
where we also propose a method to generate appropriate weight maps for the tone
fusion. Finally, an enhanced image is obtained by multiplying the equalized
illumination and enhanced reflectance. Experiments show that the proposed
algorithm produces visually pleasing output and also yields comparable
objective results to the conventional enhancement methods, while requiring
modest computational loads
Deep Semantics-Aware Photo Adjustment
Automatic photo adjustment is to mimic the photo retouching style of
professional photographers and automatically adjust photos to the learned
style. There have been many attempts to model the tone and the color adjustment
globally with low-level color statistics. Also, spatially varying photo
adjustment methods have been studied by exploiting high-level features and
semantic label maps. Those methods are semantics-aware since the color mapping
is dependent on the high-level semantic context. However, their performance is
limited to the pre-computed hand-crafted features and it is hard to reflect
user's preference to the adjustment. In this paper, we propose a deep neural
network that models the semantics-aware photo adjustment. The proposed network
exploits bilinear models that are the multiplicative interaction of the color
and the contexual features. As the contextual features we propose the semantic
adjustment map, which discovers the inherent photo retouching presets that are
applied according to the scene context. The proposed method is trained using a
robust loss with a scene parsing task. The experimental results show that the
proposed method outperforms the existing method both quantitatively and
qualitatively. The proposed method also provides users a way to retouch the
photo by their own likings by giving customized adjustment maps
A Bio-Inspired Multi-Exposure Fusion Framework for Low-light Image Enhancement
Low-light images are not conducive to human observation and computer vision
algorithms due to their low visibility. Although many image enhancement
techniques have been proposed to solve this problem, existing methods
inevitably introduce contrast under- and over-enhancement. Inspired by human
visual system, we design a multi-exposure fusion framework for low-light image
enhancement. Based on the framework, we propose a dual-exposure fusion
algorithm to provide an accurate contrast and lightness enhancement.
Specifically, we first design the weight matrix for image fusion using
illumination estimation techniques. Then we introduce our camera response model
to synthesize multi-exposure images. Next, we find the best exposure ratio so
that the synthetic image is well-exposed in the regions where the original
image is under-exposed. Finally, the enhanced result is obtained by fusing the
input image and the synthetic image according to the weight matrix. Experiments
show that our method can obtain results with less contrast and lightness
distortion compared to that of several state-of-the-art methods.Comment: Project website: https://baidut.github.io/BIMEF
MEMC-Net: Motion Estimation and Motion Compensation Driven Neural Network for Video Interpolation and Enhancement
Motion estimation (ME) and motion compensation (MC) have been widely used for
classical video frame interpolation systems over the past decades. Recently, a
number of data-driven frame interpolation methods based on convolutional neural
networks have been proposed. However, existing learning based methods typically
estimate either flow or compensation kernels, thereby limiting performance on
both computational efficiency and interpolation accuracy. In this work, we
propose a motion estimation and compensation driven neural network for video
frame interpolation. A novel adaptive warping layer is developed to integrate
both optical flow and interpolation kernels to synthesize target frame pixels.
This layer is fully differentiable such that both the flow and kernel
estimation networks can be optimized jointly. The proposed model benefits from
the advantages of motion estimation and compensation methods without using
hand-crafted features. Compared to existing methods, our approach is
computationally efficient and able to generate more visually appealing results.
Furthermore, the proposed MEMC-Net can be seamlessly adapted to several video
enhancement tasks, e.g., super-resolution, denoising, and deblocking. Extensive
quantitative and qualitative evaluations demonstrate that the proposed method
performs favorably against the state-of-the-art video frame interpolation and
enhancement algorithms on a wide range of datasets.Comment: To appear in IEEE Transactions on Pattern Analysis and Machine
Intelligenc
Simultaneous Inpainting and Denoising by Directional Global Three-part Decomposition: Connecting Variational and Fourier Domain Based Image Processing
We consider the very challenging task of restoring images (i) which have a
large number of missing pixels, (ii) whose existing pixels are corrupted by
noise and (iii) the ideal image to be restored contains both cartoon and
texture elements. The combination of these three properties makes this inverse
problem a very difficult one. The solution proposed in this manuscript is based
on directional global three-part decomposition (DG3PD) [ThaiGottschlich2016]
with directional total variation norm, directional G-norm and
-norm in curvelet domain as key ingredients of the model. Image
decomposition by DG3PD enables a decoupled inpainting and denoising of the
cartoon and texture components. A comparison to existing approaches for
inpainting and denoising shows the advantages of the proposed method. Moreover,
we regard the image restoration problem from the viewpoint of a Bayesian
framework and we discuss the connections between the proposed solution by
function space and related image representation by harmonic analysis and
pyramid decomposition
GRAINS: Generative Recursive Autoencoders for INdoor Scenes
We present a generative neural network which enables us to generate plausible
3D indoor scenes in large quantities and varieties, easily and highly
efficiently. Our key observation is that indoor scene structures are inherently
hierarchical. Hence, our network is not convolutional; it is a recursive neural
network or RvNN. Using a dataset of annotated scene hierarchies, we train a
variational recursive autoencoder, or RvNN-VAE, which performs scene object
grouping during its encoding phase and scene generation during decoding.
Specifically, a set of encoders are recursively applied to group 3D objects
based on support, surround, and co-occurrence relations in a scene, encoding
information about object spatial properties, semantics, and their relative
positioning with respect to other objects in the hierarchy. By training a
variational autoencoder (VAE), the resulting fixed-length codes roughly follow
a Gaussian distribution. A novel 3D scene can be generated hierarchically by
the decoder from a randomly sampled code from the learned distribution. We coin
our method GRAINS, for Generative Recursive Autoencoders for INdoor Scenes. We
demonstrate the capability of GRAINS to generate plausible and diverse 3D
indoor scenes and compare with existing methods for 3D scene synthesis. We show
applications of GRAINS including 3D scene modeling from 2D layouts, scene
editing, and semantic scene segmentation via PointNet whose performance is
boosted by the large quantity and variety of 3D scenes generated by our method.Comment: 21 pages, 26 figure
- …