6,813 research outputs found

    Unsupervised three-dimensional reconstruction of small rocks from a single two-dimensional image

    Get PDF
    Surfaces covered with pebbles and small rocks can often be found in nature or in human shaped environments. Generating an accurate three-dimensional model of those kind of surfaces from a reference image can be challenging, especially if one wants to be able to animate each pebble individually. To undertake this kind of job manually is time consuming and impossible to achieve in dynamic terrains animations. The method described in this paper allows unsupervised automatic generation of three-dimensional textured rocks from a two-dimensional image aiming to closely match the original image as much as possible

    Recovering Faces from Portraits with Auxiliary Facial Attributes

    Full text link
    Recovering a photorealistic face from an artistic portrait is a challenging task since crucial facial details are often distorted or completely lost in artistic compositions. To handle this loss, we propose an Attribute-guided Face Recovery from Portraits (AFRP) that utilizes a Face Recovery Network (FRN) and a Discriminative Network (DN). FRN consists of an autoencoder with residual block-embedded skip-connections and incorporates facial attribute vectors into the feature maps of input portraits at the bottleneck of the autoencoder. DN has multiple convolutional and fully-connected layers, and its role is to enforce FRN to generate authentic face images with corresponding facial attributes dictated by the input attribute vectors. %Leveraging on the spatial transformer networks, FRN automatically compensates for misalignments of portraits. % and generates aligned face images. For the preservation of identities, we impose the recovered and ground-truth faces to share similar visual features. Specifically, DN determines whether the recovered image looks like a real face and checks if the facial attributes extracted from the recovered image are consistent with given attributes. %Our method can recover high-quality photorealistic faces from unaligned portraits while preserving the identity of the face images as well as it can reconstruct a photorealistic face image with a desired set of attributes. Our method can recover photorealistic identity-preserving faces with desired attributes from unseen stylized portraits, artistic paintings, and hand-drawn sketches. On large-scale synthesized and sketch datasets, we demonstrate that our face recovery method achieves state-of-the-art results.Comment: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV

    A framework for realistic 3D tele-immersion

    Get PDF
    Meeting, socializing and conversing online with a group of people using teleconferencing systems is still quite differ- ent from the experience of meeting face to face. We are abruptly aware that we are online and that the people we are engaging with are not in close proximity. Analogous to how talking on the telephone does not replicate the experi- ence of talking in person. Several causes for these differences have been identified and we propose inspiring and innova- tive solutions to these hurdles in attempt to provide a more realistic, believable and engaging online conversational expe- rience. We present the distributed and scalable framework REVERIE that provides a balanced mix of these solutions. Applications build on top of the REVERIE framework will be able to provide interactive, immersive, photo-realistic ex- periences to a multitude of users that for them will feel much more similar to having face to face meetings than the expe- rience offered by conventional teleconferencing systems

    Variable Resolution & Dimensional Mapping For 3d Model Optimization

    Get PDF
    Three-dimensional computer models, especially geospatial architectural data sets, can be visualized in the same way humans experience the world, providing a realistic, interactive experience. Scene familiarization, architectural analysis, scientific visualization, and many other applications would benefit from finely detailed, high resolution, 3D models. Automated methods to construct these 3D models traditionally has produced data sets that are often low fidelity or inaccurate; otherwise, they are initially highly detailed, but are very labor and time intensive to construct. Such data sets are often not practical for common real-time usage and are not easily updated. This thesis proposes Variable Resolution & Dimensional Mapping (VRDM), a methodology that has been developed to address some of the limitations of existing approaches to model construction from images. Key components of VRDM are texture palettes, which enable variable and ultra-high resolution images to be easily composited; texture features, which allow image features to integrated as image or geometry, and have the ability to modify the geometric model structure to add detail. These components support a primary VRDM objective of facilitating model refinement with additional data. This can be done until the desired fidelity is achieved as practical limits of infinite detail are approached. Texture Levels, the third component, enable real-time interaction with a very detailed model, along with the flexibility of having alternate pixel data for a given area of the model and this is achieved through extra dimensions. Together these techniques have been used to construct models that can contain GBs of imagery data

    GAN-Based Differential Private Image Privacy Protection Framework for the Internet of Multimedia Things.

    Full text link
    With the development of the Internet of Multimedia Things (IoMT), an increasing amount of image data is collected by various multimedia devices, such as smartphones, cameras, and drones. This massive number of images are widely used in each field of IoMT, which presents substantial challenges for privacy preservation. In this paper, we propose a new image privacy protection framework in an effort to protect the sensitive personal information contained in images collected by IoMT devices. We aim to use deep neural network techniques to identify the privacy-sensitive content in images, and then protect it with the synthetic content generated by generative adversarial networks (GANs) with differential privacy (DP). Our experiment results show that the proposed framework can effectively protect users' privacy while maintaining image utility

    The ALHAMBRA Survey: Bayesian Photometric Redshifts with 23 bands for 3 squared degrees

    Full text link
    The ALHAMBRA (Advance Large Homogeneous Area Medium Band Redshift Astronomical) survey has observed 8 different regions of the sky, including sections of the COSMOS, DEEP2, ELAIS, GOODS-N, SDSS and Groth fields using a new photometric system with 20 contiguous ~ 300A˚300\AA filters covering the optical range, combining them with deep JHKsJHKs imaging. The observations, carried out with the Calar Alto 3.5m telescope using the wide field (0.25 sq. deg FOV) optical camera LAICA and the NIR instrument Omega-2000, correspond to ~700hrs on-target science images. The photometric system was designed to maximize the effective depth of the survey in terms of accurate spectral-type and photo-zs estimation along with the capability of identification of relatively faint emission lines. Here we present multicolor photometry and photo-zs for ~438k galaxies, detected in synthetic F814W images, complete down to I~24.5 AB, taking into account realistic noise estimates, and correcting by PSF and aperture effects with the ColorPro software. The photometric ZP have been calibrated using stellar transformation equations and refined internally, using a new technique based on the highly robust photometric redshifts measured for emission line galaxies. We calculate photometric redshifts with the BPZ2 code, which includes new empirically calibrated templates and priors. Our photo-zs have a precision of dz/(1+zs)=1dz/(1+z_s)=1% for I<22.5 and 1.4% for 22.5<I<24.5. Precisions of less than 0.5% are reached for the brighter spectroscopic sample, showing the potential of medium-band photometric surveys. The global P(z)P(z) shows a mean redshift =0.56 for I=0.86 for I<24.5 AB. The data presented here covers an effective area of 2.79 sq. deg, split into 14 strips of 58.5'x15.5' and represents ~32 hrs of on-target.Comment: The catalog data and a full resolution version of this paper is available at https://cloud.iaa.csic.es/alhambra
    • 

    corecore