168 research outputs found
Fast Deep Multi-patch Hierarchical Network for Nonhomogeneous Image Dehazing
Recently, CNN based end-to-end deep learning methods achieve superiority in
Image Dehazing but they tend to fail drastically in Non-homogeneous dehazing.
Apart from that, existing popular Multi-scale approaches are runtime intensive
and memory inefficient. In this context, we proposed a fast Deep Multi-patch
Hierarchical Network to restore Non-homogeneous hazed images by aggregating
features from multiple image patches from different spatial sections of the
hazed image with fewer number of network parameters. Our proposed method is
quite robust for different environments with various density of the haze or fog
in the scene and very lightweight as the total size of the model is around 21.7
MB. It also provides faster runtime compared to current multi-scale methods
with an average runtime of 0.0145s to process 1200x1600 HD quality image.
Finally, we show the superiority of this network on Dense Haze Removal to other
state-of-the-art models.Comment: CVPR Workshops Proceedings 202
Model Adaptation with Synthetic and Real Data for Semantic Dense Foggy Scene Understanding
This work addresses the problem of semantic scene understanding under dense
fog. Although considerable progress has been made in semantic scene
understanding, it is mainly related to clear-weather scenes. Extending
recognition methods to adverse weather conditions such as fog is crucial for
outdoor applications. In this paper, we propose a novel method, named
Curriculum Model Adaptation (CMAda), which gradually adapts a semantic
segmentation model from light synthetic fog to dense real fog in multiple
steps, using both synthetic and real foggy data. In addition, we present three
other main stand-alone contributions: 1) a novel method to add synthetic fog to
real, clear-weather scenes using semantic input; 2) a new fog density
estimator; 3) the Foggy Zurich dataset comprising real foggy images,
with pixel-level semantic annotations for images with dense fog. Our
experiments show that 1) our fog simulation slightly outperforms a
state-of-the-art competing simulation with respect to the task of semantic
foggy scene understanding (SFSU); 2) CMAda improves the performance of
state-of-the-art models for SFSU significantly by leveraging unlabeled real
foggy data. The datasets and code are publicly available.Comment: final version, ECCV 201
Learned Perceptual Image Enhancement
Learning a typical image enhancement pipeline involves minimization of a loss
function between enhanced and reference images. While L1 and L2 losses are
perhaps the most widely used functions for this purpose, they do not
necessarily lead to perceptually compelling results. In this paper, we show
that adding a learned no-reference image quality metric to the loss can
significantly improve enhancement operators. This metric is implemented using a
CNN (convolutional neural network) trained on a large-scale dataset labelled
with aesthetic preferences of human raters. This loss allows us to conveniently
perform back-propagation in our learning framework to simultaneously optimize
for similarity to a given ground truth reference and perceptual quality. This
perceptual loss is only used to train parameters of image processing operators,
and does not impose any extra complexity at inference time. Our experiments
demonstrate that this loss can be effective for tuning a variety of operators
such as local tone mapping and dehazing
- …