54,641 research outputs found
Multiple Linear Regression Haze-removal Model Based on Dark Channel Prior
Dark Channel Prior (DCP) is a widely recognized traditional dehazing
algorithm. However, it may fail in bright region and the brightness of the
restored image is darker than hazy image. In this paper, we propose an
effective method to optimize DCP. We build a multiple linear regression
haze-removal model based on DCP atmospheric scattering model and train this
model with RESIDE dataset, which aims to reduce the unexpected errors caused by
the rough estimations of transmission map t(x) and atmospheric light A. The
RESIDE dataset provides enough synthetic hazy images and their corresponding
groundtruth images to train and test. We compare the performances of different
dehazing algorithms in terms of two important full-reference metrics, the
peak-signal-to-noise ratio (PSNR) as well as the structural similarity index
measure (SSIM). The experiment results show that our model gets highest SSIM
value and its PSNR value is also higher than most of state-of-the-art dehazing
algorithms. Our results also overcome the weakness of DCP on real-world hazy
imagesComment: IEEE CPS (CSCI 2018 Int'l Conference
Image Dehazing using Bilinear Composition Loss Function
In this paper, we introduce a bilinear composition loss function to address
the problem of image dehazing. Previous methods in image dehazing use a
two-stage approach which first estimate the transmission map followed by clear
image estimation. The drawback of a two-stage method is that it tends to boost
local image artifacts such as noise, aliasing and blocking. This is especially
the case for heavy haze images captured with a low quality device. Our method
is based on convolutional neural networks. Unique in our method is the bilinear
composition loss function which directly model the correlations between
transmission map, clear image, and atmospheric light. This allows errors to be
back-propagated to each sub-network concurrently, while maintaining the
composition constraint to avoid overfitting of each sub-network. We evaluate
the effectiveness of our proposed method using both synthetic and real world
examples. Extensive experiments show that our method outperfoms
state-of-the-art methods especially for haze images with severe noise level and
compressions
Single Image Restoration for Participating Media Based on Prior Fusion
This paper describes a method to restore degraded images captured in a
participating media -- fog, turbid water, sand storm, etc. Differently from the
related work that only deal with a medium, we obtain generality by using an
image formation model and a fusion of new image priors. The model considers the
image color variation produced by the medium. The proposed restoration method
is based on the fusion of these priors and supported by statistics collected on
images acquired in both non-participating and participating media. The key of
the method is to fuse two complementary measures --- local contrast and color
data. The obtained results on underwater and foggy images demonstrate the
capabilities of the proposed method. Moreover, we evaluated our method using a
special dataset for which a ground-truth image is available.Comment: This paper is under consideration at Pattern Recognition Letter
Joint Defogging and Demosaicking
Image defogging is a technique used extensively for enhancing visual quality
of images in bad weather condition. Even though defogging algorithms have been
well studied, defogging performance is degraded by demosaicking artifacts and
sensor noise amplification in distant scenes. In order to improve visual
quality of restored images, we propose a novel approach to perform defogging
and demosaicking simultaneously. We conclude that better defogging performance
with fewer artifacts can be achieved when a defogging algorithm is combined
with a demosaicking algorithm simultaneously. We also demonstrate that the
proposed joint algorithm has the benefit of suppressing noise amplification in
distant scene. In addition, we validate our theoretical analysis and
observations for both synthesized datasets with ground truth fog-free images
and natural scene datasets captured in a raw format
Image Dehazing via Joint Estimation of Transmittance Map and Environmental Illumination
Haze limits the visibility of outdoor images, due to the existence of fog,
smoke and dust in the atmosphere. Image dehazing methods try to recover
haze-free image by removing the effect of haze from a given input image. In
this paper, we present an end to end system, which takes a hazy image as its
input and returns a dehazed image. The proposed method learns the mapping
between a hazy image and its corresponding transmittance map and the
environmental illumination, by using a multi-scale Convolutional Neural
Network. Although most of the time haze appears grayish in color, its color may
vary depending on the color of the environmental illumination. Very few of the
existing image dehazing methods have laid stress on its accurate estimation.
But the color of the dehazed image and the estimated transmittance depends on
the environmental illumination. Our proposed method exploits the relationship
between the transmittance values and the environmental illumination as per the
haze imaging model and estimates both of them. Qualitative and quantitative
evaluations show, the estimates are accurate enough.Comment: 6 pages, 9 figures, Presented at the Ninth International Conference
on Advances in Pattern Recognition(ICAPR), December 2017, Bengaluru, Indi
Real-world Noisy Image Denoising: A New Benchmark
Most of previous image denoising methods focus on additive white Gaussian
noise (AWGN). However,the real-world noisy image denoising problem with the
advancing of the computer vision techiniques. In order to promote the study on
this problem while implementing the concurrent real-world image denoising
datasets, we construct a new benchmark dataset which contains comprehensive
real-world noisy images of different natural scenes. These images are captured
by different cameras under different camera settings. We evaluate the different
denoising methods on our new dataset as well as previous datasets. Extensive
experimental results demonstrate that the recently proposed methods designed
specifically for realistic noise removal based on sparse or low rank theories
achieve better denoising performance and are more robust than other competing
methods, and the newly proposed dataset is more challenging. The constructed
dataset of real photographs is publicly available at
\url{https://github.com/csjunxu/PolyUDataset} for researchers to investigate
new real-world image denoising methods. We will add more analysis on the noise
statistics in the real photographs of our new dataset in the next version of
this article.Comment: 13 pages, 8 figures, 8 tables. arXiv admin note: text overlap with
arXiv:1707.01313 by other author
O-HAZE: a dehazing benchmark with real hazy and haze-free outdoor images
Haze removal or dehazing is a challenging ill-posed problem that has drawn a
significant attention in the last few years. Despite this growing interest, the
scientific community is still lacking a reference dataset to evaluate
objectively and quantitatively the performance of proposed dehazing methods.
The few datasets that are currently considered, both for assessment and
training of learning-based dehazing techniques, exclusively rely on synthetic
hazy images. To address this limitation, we introduce the first outdoor scenes
database (named O-HAZE) composed of pairs of real hazy and corresponding
haze-free images. In practice, hazy images have been captured in presence of
real haze, generated by professional haze machines, and OHAZE contains 45
different outdoor scenes depicting the same visual content recorded in
haze-free and hazy conditions, under the same illumination parameters. To
illustrate its usefulness, O-HAZE is used to compare a representative set of
state-of-the-art dehazing techniques, using traditional image quality metrics
such as PSNR, SSIM and CIEDE2000. This reveals the limitations of current
techniques, and questions some of their underlying assumptions.Comment: arXiv admin note: text overlap with arXiv:1804.0509
Generation of High Dynamic Range Illumination from a Single Image for the Enhancement of Undesirably Illuminated Images
This paper presents an algorithm that enhances undesirably illuminated images
by generating and fusing multi-level illuminations from a single image.The
input image is first decomposed into illumination and reflectance components by
using an edge-preserving smoothing filter. Then the reflectance component is
scaled up to improve the image details in bright areas. The illumination
component is scaled up and down to generate several illumination images that
correspond to certain camera exposure values different from the original. The
virtual multi-exposure illuminations are blended into an enhanced illumination,
where we also propose a method to generate appropriate weight maps for the tone
fusion. Finally, an enhanced image is obtained by multiplying the equalized
illumination and enhanced reflectance. Experiments show that the proposed
algorithm produces visually pleasing output and also yields comparable
objective results to the conventional enhancement methods, while requiring
modest computational loads
Single Image Dehazing through Improved Atmospheric Light Estimation
Image contrast enhancement for outdoor vision is important for smart car
auxiliary transport systems. The video frames captured in poor weather
conditions are often characterized by poor visibility. Most image dehazing
algorithms consider to use a hard threshold assumptions or user input to
estimate atmospheric light. However, the brightest pixels sometimes are objects
such as car lights or streetlights, especially for smart car auxiliary
transport systems. Simply using a hard threshold may cause a wrong estimation.
In this paper, we propose a single optimized image dehazing method that
estimates atmospheric light efficiently and removes haze through the estimation
of a semi-globally adaptive filter. The enhanced images are characterized with
little noise and good exposure in dark regions. The textures and edges of the
processed images are also enhanced significantly.Comment: Multimedia Tools and Applications (2015
Learn to Model Motion from Blurry Footages
It is difficult to recover the motion field from a real-world footage given a
mixture of camera shake and other photometric effects. In this paper we propose
a hybrid framework by interleaving a Convolutional Neural Network (CNN) and a
traditional optical flow energy. We first conduct a CNN architecture using a
novel learnable directional filtering layer. Such layer encodes the angle and
distance similarity matrix between blur and camera motion, which is able to
enhance the blur features of the camera-shake footages. The proposed CNNs are
then integrated into an iterative optical flow framework, which enable the
capability of modelling and solving both the blind deconvolution and the
optical flow estimation problems simultaneously. Our framework is trained
end-to-end on a synthetic dataset and yields competitive precision and
performance against the state-of-the-art approaches.Comment: Preprint of our paper accepted by Pattern Recognitio
- …