27 research outputs found
A Fully Progressive Approach to Single-Image Super-Resolution
Recent deep learning approaches to single image super-resolution have
achieved impressive results in terms of traditional error measures and
perceptual quality. However, in each case it remains challenging to achieve
high quality results for large upsampling factors. To this end, we propose a
method (ProSR) that is progressive both in architecture and training: the
network upsamples an image in intermediate steps, while the learning process is
organized from easy to hard, as is done in curriculum learning. To obtain more
photorealistic results, we design a generative adversarial network (GAN), named
ProGanSR, that follows the same progressive multi-scale design principle. This
not only allows to scale well to high upsampling factors (e.g., 8x) but
constitutes a principled multi-scale approach that increases the reconstruction
quality for all upsampling factors simultaneously. In particular ProSR ranks
2nd in terms of SSIM and 4th in terms of PSNR in the NTIRE2018 SISR challenge
[34]. Compared to the top-ranking team, our model is marginally lower, but runs
5 times faster
Coupled Depth Learning
In this paper we propose a method for estimating depth from a single image
using a coarse to fine approach. We argue that modeling the fine depth details
is easier after a coarse depth map has been computed. We express a global
(coarse) depth map of an image as a linear combination of a depth basis learned
from training examples. The depth basis captures spatial and statistical
regularities and reduces the problem of global depth estimation to the task of
predicting the input-specific coefficients in the linear combination. This is
formulated as a regression problem from a holistic representation of the image.
Crucially, the depth basis and the regression function are {\bf coupled} and
jointly optimized by our learning scheme. We demonstrate that this results in a
significant improvement in accuracy compared to direct regression of depth
pixel values or approaches learning the depth basis disjointly from the
regression function. The global depth estimate is then used as a guidance by a
local refinement method that introduces depth details that were not captured at
the global level. Experiments on the NYUv2 and KITTI datasets show that our
method outperforms the existing state-of-the-art at a considerably lower
computational cost for both training and testing.Comment: 10 pages, 3 Figures, 4 Tables with quantitative evaluation
Beyond Identity: What Information Is Stored in Biometric Face Templates?
Deeply-learned face representations enable the success of current face
recognition systems. Despite the ability of these representations to encode the
identity of an individual, recent works have shown that more information is
stored within, such as demographics, image characteristics, and social traits.
This threatens the user's privacy, since for many applications these templates
are expected to be solely used for recognition purposes. Knowing the encoded
information in face templates helps to develop bias-mitigating and
privacy-preserving face recognition technologies. This work aims to support the
development of these two branches by analysing face templates regarding 113
attributes. Experiments were conducted on two publicly available face
embeddings. For evaluating the predictability of the attributes, we trained a
massive attribute classifier that is additionally able to accurately state its
prediction confidence. This allows us to make more sophisticated statements
about the attribute predictability. The results demonstrate that up to 74
attributes can be accurately predicted from face templates. Especially
non-permanent attributes, such as age, hairstyles, haircolors, beards, and
various accessories, found to be easily-predictable. Since face recognition
systems aim to be robust against these variations, future research might build
on this work to develop more understandable privacy preserving solutions and
build robust and fair face templates.Comment: To appear in IJCB 202