6,813 research outputs found
Unsupervised three-dimensional reconstruction of small rocks from a single two-dimensional image
Surfaces covered with pebbles and small rocks can often be found in nature or in human shaped environments.
Generating an accurate three-dimensional model of those kind of surfaces from a reference image can be challenging,
especially if one wants to be able to animate each pebble individually. To undertake this kind of job manually
is time consuming and impossible to achieve in dynamic terrains animations.
The method described in this paper allows unsupervised automatic generation of three-dimensional textured rocks
from a two-dimensional image aiming to closely match the original image as much as possible
Recovering Faces from Portraits with Auxiliary Facial Attributes
Recovering a photorealistic face from an artistic portrait is a challenging
task since crucial facial details are often distorted or completely lost in
artistic compositions. To handle this loss, we propose an Attribute-guided Face
Recovery from Portraits (AFRP) that utilizes a Face Recovery Network (FRN) and
a Discriminative Network (DN). FRN consists of an autoencoder with residual
block-embedded skip-connections and incorporates facial attribute vectors into
the feature maps of input portraits at the bottleneck of the autoencoder. DN
has multiple convolutional and fully-connected layers, and its role is to
enforce FRN to generate authentic face images with corresponding facial
attributes dictated by the input attribute vectors. %Leveraging on the spatial
transformer networks, FRN automatically compensates for misalignments of
portraits. % and generates aligned face images. For the preservation of
identities, we impose the recovered and ground-truth faces to share similar
visual features. Specifically, DN determines whether the recovered image looks
like a real face and checks if the facial attributes extracted from the
recovered image are consistent with given attributes. %Our method can recover
high-quality photorealistic faces from unaligned portraits while preserving the
identity of the face images as well as it can reconstruct a photorealistic face
image with a desired set of attributes. Our method can recover photorealistic
identity-preserving faces with desired attributes from unseen stylized
portraits, artistic paintings, and hand-drawn sketches. On large-scale
synthesized and sketch datasets, we demonstrate that our face recovery method
achieves state-of-the-art results.Comment: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV
A framework for realistic 3D tele-immersion
Meeting, socializing and conversing online with a group of people using teleconferencing systems is still quite differ- ent from the experience of meeting face to face. We are abruptly aware that we are online and that the people we are engaging with are not in close proximity. Analogous to how talking on the telephone does not replicate the experi- ence of talking in person. Several causes for these differences have been identified and we propose inspiring and innova- tive solutions to these hurdles in attempt to provide a more realistic, believable and engaging online conversational expe- rience. We present the distributed and scalable framework REVERIE that provides a balanced mix of these solutions. Applications build on top of the REVERIE framework will be able to provide interactive, immersive, photo-realistic ex- periences to a multitude of users that for them will feel much more similar to having face to face meetings than the expe- rience offered by conventional teleconferencing systems
Variable Resolution & Dimensional Mapping For 3d Model Optimization
Three-dimensional computer models, especially geospatial architectural data sets, can be visualized in the same way humans experience the world, providing a realistic, interactive experience. Scene familiarization, architectural analysis, scientific visualization, and many other applications would benefit from finely detailed, high resolution, 3D models. Automated methods to construct these 3D models traditionally has produced data sets that are often low fidelity or inaccurate; otherwise, they are initially highly detailed, but are very labor and time intensive to construct. Such data sets are often not practical for common real-time usage and are not easily updated. This thesis proposes Variable Resolution & Dimensional Mapping (VRDM), a methodology that has been developed to address some of the limitations of existing approaches to model construction from images. Key components of VRDM are texture palettes, which enable variable and ultra-high resolution images to be easily composited; texture features, which allow image features to integrated as image or geometry, and have the ability to modify the geometric model structure to add detail. These components support a primary VRDM objective of facilitating model refinement with additional data. This can be done until the desired fidelity is achieved as practical limits of infinite detail are approached. Texture Levels, the third component, enable real-time interaction with a very detailed model, along with the flexibility of having alternate pixel data for a given area of the model and this is achieved through extra dimensions. Together these techniques have been used to construct models that can contain GBs of imagery data
GAN-Based Differential Private Image Privacy Protection Framework for the Internet of Multimedia Things.
With the development of the Internet of Multimedia Things (IoMT), an increasing amount of image data is collected by various multimedia devices, such as smartphones, cameras, and drones. This massive number of images are widely used in each field of IoMT, which presents substantial challenges for privacy preservation. In this paper, we propose a new image privacy protection framework in an effort to protect the sensitive personal information contained in images collected by IoMT devices. We aim to use deep neural network techniques to identify the privacy-sensitive content in images, and then protect it with the synthetic content generated by generative adversarial networks (GANs) with differential privacy (DP). Our experiment results show that the proposed framework can effectively protect users' privacy while maintaining image utility
Recommended from our members
Virtual viewpoint three-dimensional panorama
Conventional panoramic images are known to provide for an enhanced field of view in which the scene
always has a fixed appearance. The idea presented in this paper focuses on the use of the concept of virtual
viewpoint creation to generate different panoramic images of the same scene with three-dimensional
component. Three-dimensional effect in a resultant panorama is realized by superimposing a stereo-pair of
panoramic images
The ALHAMBRA Survey: Bayesian Photometric Redshifts with 23 bands for 3 squared degrees
The ALHAMBRA (Advance Large Homogeneous Area Medium Band Redshift
Astronomical) survey has observed 8 different regions of the sky, including
sections of the COSMOS, DEEP2, ELAIS, GOODS-N, SDSS and Groth fields using a
new photometric system with 20 contiguous ~ filters covering the
optical range, combining them with deep imaging. The observations,
carried out with the Calar Alto 3.5m telescope using the wide field (0.25 sq.
deg FOV) optical camera LAICA and the NIR instrument Omega-2000, correspond to
~700hrs on-target science images. The photometric system was designed to
maximize the effective depth of the survey in terms of accurate spectral-type
and photo-zs estimation along with the capability of identification of
relatively faint emission lines. Here we present multicolor photometry and
photo-zs for ~438k galaxies, detected in synthetic F814W images, complete down
to I~24.5 AB, taking into account realistic noise estimates, and correcting by
PSF and aperture effects with the ColorPro software. The photometric ZP have
been calibrated using stellar transformation equations and refined internally,
using a new technique based on the highly robust photometric redshifts measured
for emission line galaxies. We calculate photometric redshifts with the BPZ2
code, which includes new empirically calibrated templates and priors. Our
photo-zs have a precision of for I<22.5 and 1.4% for
22.5<I<24.5. Precisions of less than 0.5% are reached for the brighter
spectroscopic sample, showing the potential of medium-band photometric surveys.
The global shows a mean redshift =0.56 for I=0.86 for
I<24.5 AB. The data presented here covers an effective area of 2.79 sq. deg,
split into 14 strips of 58.5'x15.5' and represents ~32 hrs of on-target.Comment: The catalog data and a full resolution version of this paper is
available at https://cloud.iaa.csic.es/alhambra
- âŠ