1,756 research outputs found
Calipso: Physics-based Image and Video Editing through CAD Model Proxies
We present Calipso, an interactive method for editing images and videos in a
physically-coherent manner. Our main idea is to realize physics-based
manipulations by running a full physics simulation on proxy geometries given by
non-rigidly aligned CAD models. Running these simulations allows us to apply
new, unseen forces to move or deform selected objects, change physical
parameters such as mass or elasticity, or even add entire new objects that
interact with the rest of the underlying scene. In Calipso, the user makes
edits directly in 3D; these edits are processed by the simulation and then
transfered to the target 2D content using shape-to-image correspondences in a
photo-realistic rendering process. To align the CAD models, we introduce an
efficient CAD-to-image alignment procedure that jointly minimizes for rigid and
non-rigid alignment while preserving the high-level structure of the input
shape. Moreover, the user can choose to exploit image flow to estimate scene
motion, producing coherent physical behavior with ambient dynamics. We
demonstrate Calipso's physics-based editing on a wide range of examples
producing myriad physical behavior while preserving geometric and visual
consistency.Comment: 11 page
The GAN that Warped: Semantic Attribute Editing with Unpaired Data
Deep neural networks have recently been used to edit images with great success, in particular for faces.However, they are often limited to only being able to work at a restricted range of resolutions. Many methods are so flexible that face edits can often result in an unwanted loss of identity.This work proposes to learn how to perform semantic image edits through the application of smooth warp fields.Previous approaches that attempted to use warping for semantic edits required paired data, \ie example images of the same subject with different semantic attributes.In contrast, we employ recent advances in Generative Adversarial Networks that allow our model to be trained with unpaired data.We demonstrate face editing at very high resolutions (4k images) with a single forward pass of a deep network at a lower resolution.We also show that our edits are substantially better at preserving the subject's identity.The robustness of our approach is demonstrated by showing plausible image editing results on the Cub200 birds dataset.To our knowledge this has not been previously accomplished, due the challenging nature of the dataset
Make Your Brief Stroke Real and Stereoscopic: 3D-Aware Simplified Sketch to Portrait Generation
Creating the photo-realistic version of people sketched portraits is useful
to various entertainment purposes. Existing studies only generate portraits in
the 2D plane with fixed views, making the results less vivid. In this paper, we
present Stereoscopic Simplified Sketch-to-Portrait (SSSP), which explores the
possibility of creating Stereoscopic 3D-aware portraits from simple contour
sketches by involving 3D generative models. Our key insight is to design
sketch-aware constraints that can fully exploit the prior knowledge of a
tri-plane-based 3D-aware generative model. Specifically, our designed
region-aware volume rendering strategy and global consistency constraint
further enhance detail correspondences during sketch encoding. Moreover, in
order to facilitate the usage of layman users, we propose a Contour-to-Sketch
module with vector quantized representations, so that easily drawn contours can
directly guide the generation of 3D portraits. Extensive comparisons show that
our method generates high-quality results that match the sketch. Our usability
study verifies that our system is greatly preferred by user.Comment: Project Page on https://hangz-nju-cuhk.github.io
- …