20,547 research outputs found
Synthesizing Normalized Faces from Facial Identity Features
We present a method for synthesizing a frontal, neutral-expression image of a
person's face given an input face photograph. This is achieved by learning to
generate facial landmarks and textures from features extracted from a
facial-recognition network. Unlike previous approaches, our encoding feature
vector is largely invariant to lighting, pose, and facial expression.
Exploiting this invariance, we train our decoder network using only frontal,
neutral-expression photographs. Since these photographs are well aligned, we
can decompose them into a sparse set of landmark points and aligned texture
maps. The decoder then predicts landmarks and textures independently and
combines them using a differentiable image warping operation. The resulting
images can be used for a number of applications, such as analyzing facial
attributes, exposure and white balance adjustment, or creating a 3-D avatar
DSLR-Quality Photos on Mobile Devices with Deep Convolutional Networks
Despite a rapid rise in the quality of built-in smartphone cameras, their
physical limitations - small sensor size, compact lenses and the lack of
specific hardware, - impede them to achieve the quality results of DSLR
cameras. In this work we present an end-to-end deep learning approach that
bridges this gap by translating ordinary photos into DSLR-quality images. We
propose learning the translation function using a residual convolutional neural
network that improves both color rendition and image sharpness. Since the
standard mean squared loss is not well suited for measuring perceptual image
quality, we introduce a composite perceptual error function that combines
content, color and texture losses. The first two losses are defined
analytically, while the texture loss is learned in an adversarial fashion. We
also present DPED, a large-scale dataset that consists of real photos captured
from three different phones and one high-end reflex camera. Our quantitative
and qualitative assessments reveal that the enhanced image quality is
comparable to that of DSLR-taken photos, while the methodology is generalized
to any type of digital camera
CNN-SLAM: Real-time dense monocular SLAM with learned depth prediction
Given the recent advances in depth prediction from Convolutional Neural
Networks (CNNs), this paper investigates how predicted depth maps from a deep
neural network can be deployed for accurate and dense monocular reconstruction.
We propose a method where CNN-predicted dense depth maps are naturally fused
together with depth measurements obtained from direct monocular SLAM. Our
fusion scheme privileges depth prediction in image locations where monocular
SLAM approaches tend to fail, e.g. along low-textured regions, and vice-versa.
We demonstrate the use of depth prediction for estimating the absolute scale of
the reconstruction, hence overcoming one of the major limitations of monocular
SLAM. Finally, we propose a framework to efficiently fuse semantic labels,
obtained from a single frame, with dense SLAM, yielding semantically coherent
scene reconstruction from a single view. Evaluation results on two benchmark
datasets show the robustness and accuracy of our approach.Comment: 10 pages, 6 figures, IEEE Computer Society Conference on Computer
Vision and Pattern Recognition (CVPR), Hawaii, USA, June, 2017. The first two
authors contribute equally to this pape
Sparse Coral Classification Using Deep Convolutional Neural Networks
Autonomous repair of deep-sea coral reefs is a recent proposed idea to
support the oceans ecosystem in which is vital for commercial fishing, tourism
and other species. This idea can be operated through using many small
autonomous underwater vehicles (AUVs) and swarm intelligence techniques to
locate and replace chunks of coral which have been broken off, thus enabling
re-growth and maintaining the habitat. The aim of this project is developing
machine vision algorithms to enable an underwater robot to locate a coral reef
and a chunk of coral on the seabed and prompt the robot to pick it up. Although
there is no literature on this particular problem, related work on fish
counting may give some insight into the problem. The technical challenges are
principally due to the potential lack of clarity of the water and platform
stabilization as well as spurious artifacts (rocks, fish, and crabs). We
present an efficient sparse classification for coral species using supervised
deep learning method called Convolutional Neural Networks (CNNs). We compute
Weber Local Descriptor (WLD), Phase Congruency (PC), and Zero Component
Analysis (ZCA) Whitening to extract shape and texture feature descriptors,
which are employed to be supplementary channels (feature-based maps) besides
basic spatial color channels (spatial-based maps) of coral input image, we also
experiment state-of-art preprocessing underwater algorithms for image
enhancement and color normalization and color conversion adjustment. Our
proposed coral classification method is developed under MATLAB platform, and
evaluated by two different coral datasets (University of California San Diego's
Moorea Labeled Corals, and Heriot-Watt University's Atlantic Deep Sea).Comment: Thesis Submitted for the Degree of MSc Erasmus Mundus in Vision and
Robotics (VIBOT 2014
A Synergistic Approach for Recovering Occlusion-Free Textured 3D Maps of Urban Facades from Heterogeneous Cartographic Data
In this paper we present a practical approach for generating an
occlusion-free textured 3D map of urban facades by the synergistic use of
terrestrial images, 3D point clouds and area-based information. Particularly in
dense urban environments, the high presence of urban objects in front of the
facades causes significant difficulties for several stages in computational
building modeling. Major challenges lie on the one hand in extracting complete
3D facade quadrilateral delimitations and on the other hand in generating
occlusion-free facade textures. For these reasons, we describe a
straightforward approach for completing and recovering facade geometry and
textures by exploiting the data complementarity of terrestrial multi-source
imagery and area-based information
PetroSurf3D - A Dataset for high-resolution 3D Surface Segmentation
The development of powerful 3D scanning hardware and reconstruction
algorithms has strongly promoted the generation of 3D surface reconstructions
in different domains. An area of special interest for such 3D reconstructions
is the cultural heritage domain, where surface reconstructions are generated to
digitally preserve historical artifacts. While reconstruction quality nowadays
is sufficient in many cases, the robust analysis (e.g. segmentation, matching,
and classification) of reconstructed 3D data is still an open topic. In this
paper, we target the automatic and interactive segmentation of high-resolution
3D surface reconstructions from the archaeological domain. To foster research
in this field, we introduce a fully annotated and publicly available
large-scale 3D surface dataset including high-resolution meshes, depth maps and
point clouds as a novel benchmark dataset to the community. We provide baseline
results for our existing random forest-based approach and for the first time
investigate segmentation with convolutional neural networks (CNNs) on the data.
Results show that both approaches have complementary strengths and weaknesses
and that the provided dataset represents a challenge for future research.Comment: CBMI Submission; Dataset and more information can be found at
http://lrs.icg.tugraz.at/research/petroglyphsegmentation
- …