75,850 research outputs found
Spatial relationship based scene analysis and synthesis
In this thesis, we propose a new representation, which we name Interaction Bisector
Surface (IBS), that can describe the general nature of spatial relationship. We show
that the IBS can be applied in 3D scene analysis, retrieval and synthesis.
Despite the fact that the spatial relationship between different objects plays a significant
role in describing the context, few works have focused on elaborating a representation
that can describe arbitrary interactions between different objects. Previous
methods simply concatenate the individual state vectors to produce a joint space, or
only use simple representations such as relative vectors or contacts to describe the context.
Such representations do not contain detailed information of spatial relationships.
They cannot describe complex interactions such as hooking and enclosure.
The IBS is a data structure with rich information about the interaction. It provides
the topological, geometric and correspondence features that can be used to classify and
recognize interactions. The topological features are at the most abstract level and it can
be used to recognize spatial relationships such as enclosure, hooking and surrounding.
The geometric features encode the fine details of interactions. The correspondence
feature describes which parts of the scene elements contribute to the interaction and
is especially useful for recognizing character-object interactions. We show examples
of successful classification and retrieval of different types of data including indoor
static scenes and dynamic scenes which contain character-object interactions. We also
conduct an exhaustive comparison which shows that our method outperforms existing
approaches.
We also propose a novel approach to automatically synthesizing new interactions
from example scenes and new objects. Given an example scene composed of two objects,
the open space between the objects is abstracted by the IBS. Then, an translation,
rotation and scale equivariant feature called shape coverage feature, which encodes
how the point in the open space is surrounded by the environment, is computed near
the IBS and around the open space of the new objects. Finally, a novel scene is synthesized
by conducting a partial matching of the open space around the new objects with
the IBS. Using our approach, new scenes can be automatically synthesized from example
scenes and new objects without relying on label information, which is especially
useful when the data of scenes and objects come from multiple sources
The effectiveness of DNA databases in relation to their purpose and content : a systematic review
Different stakeholders use forensic DNA databases for different purposes; for example, law enforcement agencies use them as an investigative tool to identify suspects, and criminologists use them to study the offending patterns of unidentified suspects. A number of researchers have already studied their effectiveness, but none has performed an overview of the relevant literature. Such an overview could help future researchers and policymakers by evaluating their creation, use and expansion. Using a systematic review, this article synthesizes the most relevant research into the effectiveness of forensic DNA databases published between January 1985 and March 2018. We report the results of the selected studies and look deeper into the evidence by evaluating the relationship between the purpose, content, and effectiveness of DNA databases, three inseparable elements in this type of research. We classify the studies by purposes: (i) detection and clearance; (ii) deterrence; and (iii) criminological scientific knowledge. Each category uses different measurements to evaluate effectiveness. The majority of these studies report positive results, supporting the assumption that DNA databases are an effective tool for the police, society, and criminologists. (C) 2019 Elsevier B.V. All rights reserved
Visual Importance-Biased Image Synthesis Animation
Present ray tracing algorithms are computationally intensive, requiring hours of computing time for complex scenes. Our previous work has dealt with the development of an overall approach to the application of visual attention to progressive and adaptive ray-tracing techniques. The approach facilitates large computational savings by modulating the supersampling rates in an image by the visual importance of the region being rendered. This paper extends the approach by incorporating temporal changes into the models and techniques developed, as it is expected that further efficiency savings can be reaped for animated scenes. Applications for this approach include entertainment, visualisation and simulation
Hierarchy Composition GAN for High-fidelity Image Synthesis
Despite the rapid progress of generative adversarial networks (GANs) in image
synthesis in recent years, the existing image synthesis approaches work in
either geometry domain or appearance domain alone which often introduces
various synthesis artifacts. This paper presents an innovative Hierarchical
Composition GAN (HIC-GAN) that incorporates image synthesis in geometry and
appearance domains into an end-to-end trainable network and achieves superior
synthesis realism in both domains simultaneously. We design an innovative
hierarchical composition mechanism that is capable of learning realistic
composition geometry and handling occlusions while multiple foreground objects
are involved in image composition. In addition, we introduce a novel attention
mask mechanism that guides to adapt the appearance of foreground objects which
also helps to provide better training reference for learning in geometry
domain. Extensive experiments on scene text image synthesis, portrait editing
and indoor rendering tasks show that the proposed HIC-GAN achieves superior
synthesis performance qualitatively and quantitatively.Comment: 11 pages, 8 figure
Learning to Synthesize a 4D RGBD Light Field from a Single Image
We present a machine learning algorithm that takes as input a 2D RGB image
and synthesizes a 4D RGBD light field (color and depth of the scene in each ray
direction). For training, we introduce the largest public light field dataset,
consisting of over 3300 plenoptic camera light fields of scenes containing
flowers and plants. Our synthesis pipeline consists of a convolutional neural
network (CNN) that estimates scene geometry, a stage that renders a Lambertian
light field using that geometry, and a second CNN that predicts occluded rays
and non-Lambertian effects. Our algorithm builds on recent view synthesis
methods, but is unique in predicting RGBD for each light field ray and
improving unsupervised single image depth estimation by enforcing consistency
of ray depths that should intersect the same scene point. Please see our
supplementary video at https://youtu.be/yLCvWoQLnmsComment: International Conference on Computer Vision (ICCV) 201
- …