149 research outputs found
Mean value coordinates–based caricature and expression synthesis
We present a novel method for caricature synthesis based on mean value coordinates (MVC). Our method can be applied to any single frontal face image to learn a specified caricature face pair for frontal and 3D caricature synthesis. This technique only requires one or a small number of exemplar pairs and a natural frontal face image training set, while the system can transfer the style of the exemplar pair across individuals. Further exaggeration can be fulfilled in a controllable way. Our method is further applied to facial expression transfer, interpolation, and exaggeration, which are applications of expression editing. Additionally, we have extended our approach to 3D caricature synthesis based on the 3D version of MVC. With experiments we demonstrate that the transferred expressions are credible and the resulting caricatures can be characterized and recognized
Alive Caricature from 2D to 3D
Caricature is an art form that expresses subjects in abstract, simple and
exaggerated view. While many caricatures are 2D images, this paper presents an
algorithm for creating expressive 3D caricatures from 2D caricature images with
a minimum of user interaction. The key idea of our approach is to introduce an
intrinsic deformation representation that has a capacity of extrapolation
enabling us to create a deformation space from standard face dataset, which
maintains face constraints and meanwhile is sufficiently large for producing
exaggerated face models. Built upon the proposed deformation representation, an
optimization model is formulated to find the 3D caricature that captures the
style of the 2D caricature image automatically. The experiments show that our
approach has better capability in expressing caricatures than those fitting
approaches directly using classical parametric face models such as 3DMM and
FaceWareHouse. Moreover, our approach is based on standard face datasets and
avoids constructing complicated 3D caricature training set, which provides
great flexibility in real applications.Comment: Accepted to CVPR 201
Modeling Caricature Expressions by 3D Blendshape and Dynamic Texture
The problem of deforming an artist-drawn caricature according to a given
normal face expression is of interest in applications such as social media,
animation and entertainment. This paper presents a solution to the problem,
with an emphasis on enhancing the ability to create desired expressions and
meanwhile preserve the identity exaggeration style of the caricature, which
imposes challenges due to the complicated nature of caricatures. The key of our
solution is a novel method to model caricature expression, which extends
traditional 3DMM representation to caricature domain. The method consists of
shape modelling and texture generation for caricatures. Geometric optimization
is developed to create identity-preserving blendshapes for reconstructing
accurate and stable geometric shape, and a conditional generative adversarial
network (cGAN) is designed for generating dynamic textures under target
expressions. The combination of both shape and texture components makes the
non-trivial expressions of a caricature be effectively defined by the extension
of the popular 3DMM representation and a caricature can thus be flexibly
deformed into arbitrary expressions with good results visually in both shape
and color spaces. The experiments demonstrate the effectiveness of the proposed
method.Comment: Accepted by the 28th ACM International Conference on Multimedia (ACM
MM 2020
- …