6,704 research outputs found
On the Use of Deep Learning for Blind Image Quality Assessment
In this work we investigate the use of deep learning for distortion-generic
blind image quality assessment. We report on different design choices, ranging
from the use of features extracted from pre-trained Convolutional Neural
Networks (CNNs) as a generic image description, to the use of features
extracted from a CNN fine-tuned for the image quality task. Our best proposal,
named DeepBIQ, estimates the image quality by average pooling the scores
predicted on multiple sub-regions of the original image. The score of each
sub-region is computed using a Support Vector Regression (SVR) machine taking
as input features extracted using a CNN fine-tuned for category-based image
quality assessment. Experimental results on the LIVE In the Wild Image Quality
Challenge Database and on the LIVE Image Quality Assessment Database show that
DeepBIQ outperforms the state-of-the-art methods compared, having a Linear
Correlation Coefficient (LCC) with human subjective scores of almost 0.91 and
0.98 respectively. Furthermore, in most of the cases, the quality score
predictions of DeepBIQ are closer to the average observer than those of a
generic human observer
Learning Digital Camera Pipeline for Extreme Low-Light Imaging
In low-light conditions, a conventional camera imaging pipeline produces
sub-optimal images that are usually dark and noisy due to a low photon count
and low signal-to-noise ratio (SNR). We present a data-driven approach that
learns the desired properties of well-exposed images and reflects them in
images that are captured in extremely low ambient light environments, thereby
significantly improving the visual quality of these low-light images. We
propose a new loss function that exploits the characteristics of both
pixel-wise and perceptual metrics, enabling our deep neural network to learn
the camera processing pipeline to transform the short-exposure, low-light RAW
sensor data to well-exposed sRGB images. The results show that our method
outperforms the state-of-the-art according to psychophysical tests as well as
pixel-wise standard metrics and recent learning-based perceptual image quality
measures
NTIRE 2020 Challenge on NonHomogeneous Dehazing
This paper reviews the NTIRE 2020 Challenge on NonHomogeneous Dehazing of
images (restoration of rich details in hazy image). We focus on the proposed
solutions and their results evaluated on NH-Haze, a novel dataset consisting of
55 pairs of real haze free and nonhomogeneous hazy images recorded outdoor.
NH-Haze is the first realistic nonhomogeneous haze dataset that provides ground
truth images. The nonhomogeneous haze has been produced using a professional
haze generator that imitates the real conditions of haze scenes. 168
participants registered in the challenge and 27 teams competed in the final
testing phase. The proposed solutions gauge the state-of-the-art in image
dehazing.Comment: CVPR Workshops Proceedings 202
Dense Haze: A benchmark for image dehazing with dense-haze and haze-free images
Single image dehazing is an ill-posed problem that has recently drawn
important attention. Despite the significant increase in interest shown for
dehazing over the past few years, the validation of the dehazing methods
remains largely unsatisfactory, due to the lack of pairs of real hazy and
corresponding haze-free reference images. To address this limitation, we
introduce Dense-Haze - a novel dehazing dataset. Characterized by dense and
homogeneous hazy scenes, Dense-Haze contains 33 pairs of real hazy and
corresponding haze-free images of various outdoor scenes. The hazy scenes have
been recorded by introducing real haze, generated by professional haze
machines. The hazy and haze-free corresponding scenes contain the same visual
content captured under the same illumination parameters. Dense-Haze dataset
aims to push significantly the state-of-the-art in single-image dehazing by
promoting robust methods for real and various hazy scenes. We also provide a
comprehensive qualitative and quantitative evaluation of state-of-the-art
single image dehazing techniques based on the Dense-Haze dataset. Not
surprisingly, our study reveals that the existing dehazing techniques perform
poorly for dense homogeneous hazy scenes and that there is still much room for
improvement.Comment: 5 pages, 2 figure
Fully-automatic inverse tone mapping algorithm based on dynamic mid-level tone mapping
High Dynamic Range (HDR) displays can show images with higher color contrast levels and peak luminosities than the common Low Dynamic Range (LDR) displays. However, most existing video content is recorded and/or graded in LDR format. To show LDR content on HDR displays, it needs to be up-scaled using a so-called inverse tone mapping algorithm. Several techniques for inverse tone mapping have been proposed in the last years, going from simple approaches based on global and local operators to more advanced algorithms such as neural networks. Some of the drawbacks of existing techniques for inverse tone mapping are the need for human intervention, the high computation time for more advanced algorithms, limited low peak brightness, and the lack of the preservation of the artistic intentions. In this paper, we propose a fully-automatic inverse tone mapping operator based on mid-level mapping capable of real-time video processing. Our proposed algorithm allows expanding LDR images into HDR images with peak brightness over 1000 nits, preserving the artistic intentions inherent to the HDR domain. We assessed our results using the full-reference objective quality metrics HDR-VDP-2.2 and DRIM, and carrying out a subjective pair-wise comparison experiment. We compared our results with those obtained with the most recent methods found in the literature. Experimental results demonstrate that our proposed method outperforms the current state-of-the-art of simple inverse tone mapping methods and its performance is similar to other more complex and time-consuming advanced techniques
A Bio-Inspired Multi-Exposure Fusion Framework for Low-light Image Enhancement
Low-light images are not conducive to human observation and computer vision
algorithms due to their low visibility. Although many image enhancement
techniques have been proposed to solve this problem, existing methods
inevitably introduce contrast under- and over-enhancement. Inspired by human
visual system, we design a multi-exposure fusion framework for low-light image
enhancement. Based on the framework, we propose a dual-exposure fusion
algorithm to provide an accurate contrast and lightness enhancement.
Specifically, we first design the weight matrix for image fusion using
illumination estimation techniques. Then we introduce our camera response model
to synthesize multi-exposure images. Next, we find the best exposure ratio so
that the synthetic image is well-exposed in the regions where the original
image is under-exposed. Finally, the enhanced result is obtained by fusing the
input image and the synthetic image according to the weight matrix. Experiments
show that our method can obtain results with less contrast and lightness
distortion compared to that of several state-of-the-art methods.Comment: Project website: https://baidut.github.io/BIMEF
O-HAZE: a dehazing benchmark with real hazy and haze-free outdoor images
Haze removal or dehazing is a challenging ill-posed problem that has drawn a
significant attention in the last few years. Despite this growing interest, the
scientific community is still lacking a reference dataset to evaluate
objectively and quantitatively the performance of proposed dehazing methods.
The few datasets that are currently considered, both for assessment and
training of learning-based dehazing techniques, exclusively rely on synthetic
hazy images. To address this limitation, we introduce the first outdoor scenes
database (named O-HAZE) composed of pairs of real hazy and corresponding
haze-free images. In practice, hazy images have been captured in presence of
real haze, generated by professional haze machines, and OHAZE contains 45
different outdoor scenes depicting the same visual content recorded in
haze-free and hazy conditions, under the same illumination parameters. To
illustrate its usefulness, O-HAZE is used to compare a representative set of
state-of-the-art dehazing techniques, using traditional image quality metrics
such as PSNR, SSIM and CIEDE2000. This reveals the limitations of current
techniques, and questions some of their underlying assumptions.Comment: arXiv admin note: text overlap with arXiv:1804.0509
Quantitative Evaluation of Base and Detail Decomposition Filters Based on their Artifacts
This paper introduces a quantitative evaluation of filters that seek to
separate an image into its large-scale variations, the base layer, and its
fine-scale variations, the detail layer. Such methods have proliferated with
the development of HDR imaging and the proposition of many new tone-mapping
operators. We argue that an objective quality measurement for all methods can
be based on their artifacts. To this aim, the four main recurrent artifacts are
described and mathematically characterized. Among them two are classic, the
luminance halo and the staircase effect, but we show the relevance of two more,
the contrast halo and the compartmentalization effect. For each of these
artifacts we design a test-pattern and its attached measurement formula. Then
we fuse these measurements into a single quality mark, and obtain in that way a
ranking method valid for all filters performing a base+detail decomposition.
This synthetic ranking is applied to seven filters representative of the
literature and shown to agree with expert artifact rejection criteria.Comment: 12 pages; 11 figures; 2 tables; supplementary material available
(link given in the paper
Personality and cognitive factors in the assessment of multimodal stimuli in immersive virtual environments
Literature in the study of human response to immersive virtual reality systems often deals with the phenomenon of presence. It can be shown that audio and imagery with spatial information can interact to affect presence in users of immersive virtual reality. It has also been shown that there is variation between individuals in the experience of presence in VR. The relationship between these effects has hitherto not been fully explored. This thesis aims to identify and evaluate the relation- ships between spatial audio rendering and spatial relationships between audio and visual objects and cognitive and personality differences which account for variation in the experience of presence in VR with spatial audio. This thesis compares mea- sures of audiovisual quality of experience with an existing model of presence in a factor-analytical paradigm. Scores on these dimensions were compared between en- vironments which are similar or dissimilar to pre-exposure conditions and compared between when participants believed they were listening to real-world or headphone rendered audio events. Differences between audiovisual treatments, including au- dio rendering methods and audiovisual spatial relationships, were compared with differences attributed to cognitive and personality factors identified as significant predictors using hierarchical modelling. It was found that audiovisual quality of experience relates to subscales of presence by being independent of reported visual realism and involvement, but combines linearly with these factors to contribute to ’spatial presence’, a dimension of overall presence which is identified as the largest component in the construct. It was also found that, although manipulation of the spatial information content of audiovisual stimuli was a predictor of audiovisual quality of experience, this effect is overshadowed by inter-participant variation. In- teractive effects between extraversion, empathy, ease of resolving visual detail, and systematisation and are better predictors of quality of experience and spatial pres- ence than the changes to spatial information content investigated in this work. An- choring biases are also identified which suggest that novel environments are rated higher on audiovisual quality than those geometrically similar to the pre-exposure environment. These findings constitute support for a novel framework for assessing propensity for presence in terms of an information-processing model
Scene Segmentation-Based Luminance Adjustment for Multi-Exposure Image Fusion
We propose a novel method for adjusting luminance for multi-exposure image
fusion. For the adjustment, two novel scene segmentation approaches based on
luminance distribution are also proposed. Multi-exposure image fusion is a
method for producing images that are expected to be more informative and
perceptually appealing than any of the input ones, by directly fusing photos
taken with different exposures. However, existing fusion methods often produce
unclear fused images when input images do not have a sufficient number of
different exposure levels. In this paper, we point out that adjusting the
luminance of input images makes it possible to improve the quality of the final
fused images. This insight is the basis of the proposed method. The proposed
method enables us to produce high-quality images, even when undesirable inputs
are given. Visual comparison results show that the proposed method can produce
images that clearly represent a whole scene. In addition, multi-exposure image
fusion with the proposed method outperforms state-of-the-art fusion methods in
terms of MEF-SSIM, discrete entropy, tone mapped image quality index, and
statistical naturalness.Comment: will be published in IEEE Transactions on Image Processin
- …