75 research outputs found
User Constrained Thumbnail Generation using Adaptive Convolutions
Thumbnails are widely used all over the world as a preview for digital
images. In this work we propose a deep neural framework to generate thumbnails
of any size and aspect ratio, even for unseen values during training, with high
accuracy and precision. We use Global Context Aggregation (GCA) and a modified
Region Proposal Network (RPN) with adaptive convolutions to generate thumbnails
in real time. GCA is used to selectively attend and aggregate the global
context information from the entire image while the RPN is used to predict
candidate bounding boxes for the thumbnail image. Adaptive convolution
eliminates the problem of generating thumbnails of various aspect ratios by
using filter weights dynamically generated from the aspect ratio information.
The experimental results indicate the superior performance of the proposed
model over existing state-of-the-art techniques.Comment: International Conference on Acoustics, Speech, and Signal
Processing(ICASSP), 201
Image Aesthetics Assessment Using Composite Features from off-the-Shelf Deep Models
Deep convolutional neural networks have recently achieved great success on
image aesthetics assessment task. In this paper, we propose an efficient method
which takes the global, local and scene-aware information of images into
consideration and exploits the composite features extracted from corresponding
pretrained deep learning models to classify the derived features with support
vector machine. Contrary to popular methods that require fine-tuning or
training a new model from scratch, our training-free method directly takes the
deep features generated by off-the-shelf models for image classification and
scene recognition. Also, we analyzed the factors that could influence the
performance from two aspects: the architecture of the deep neural network and
the contribution of local and scene-aware information. It turns out that deep
residual network could produce more aesthetics-aware image representation and
composite features lead to the improvement of overall performance. Experiments
on common large-scale aesthetics assessment benchmarks demonstrate that our
method outperforms the state-of-the-art results in photo aesthetics assessment.Comment: Accepted by ICIP 201
An Universal Image Attractiveness Ranking Framework
We propose a new framework to rank image attractiveness using a novel
pairwise deep network trained with a large set of side-by-side multi-labeled
image pairs from a web image index. The judges only provide relative ranking
between two images without the need to directly assign an absolute score, or
rate any predefined image attribute, thus making the rating more intuitive and
accurate. We investigate a deep attractiveness rank net (DARN), a combination
of deep convolutional neural network and rank net, to directly learn an
attractiveness score mean and variance for each image and the underlying
criteria the judges use to label each pair. The extension of this model
(DARN-V2) is able to adapt to individual judge's personal preference. We also
show the attractiveness of search results are significantly improved by using
this attractiveness information in a real commercial search engine. We evaluate
our model against other state-of-the-art models on our side-by-side web test
data and another public aesthetic data set. With much less judgments (1M vs
50M), our model outperforms on side-by-side labeled data, and is comparable on
data labeled by absolute score.Comment: Accepted by 2019 Winter Conference on Application of Computer Vision
(WACV
No-reference Image Denoising Quality Assessment
A wide variety of image denoising methods are available now. However, the
performance of a denoising algorithm often depends on individual input noisy
images as well as its parameter setting. In this paper, we present a
no-reference image denoising quality assessment method that can be used to
select for an input noisy image the right denoising algorithm with the optimal
parameter setting. This is a challenging task as no ground truth is available.
This paper presents a data-driven approach to learn to predict image denoising
quality. Our method is based on the observation that while individual existing
quality metrics and denoising models alone cannot robustly rank denoising
results, they often complement each other. We accordingly design denoising
quality features based on these existing metrics and models and then use Random
Forests Regression to aggregate them into a more powerful unified metric. Our
experiments on images with various types and levels of noise show that our
no-reference denoising quality assessment method significantly outperforms the
state-of-the-art quality metrics. This paper also provides a method that
leverages our quality assessment method to automatically tune the parameter
settings of a denoising algorithm for an input noisy image to produce an
optimal denoising result.Comment: 17 pages, 41 figures, accepted by Computer Vision Conference (CVC)
201
- …