655 research outputs found
Surface parameterization over regular domains
Surface parameterization has been widely studied and it has been playing a critical role in many geometric processing tasks in graphics, computer-aided design, visualization, vision, physical simulation and etc. Regular domains, such as polycubes, are favored due to their structural regularity and geometric simplicity. This thesis focuses on studying the surface parameterization over regular domains, i.e. polycubes, and develops effective computation algorithms. Firstly, the motivation for surface parameterization and polycube mapping is introduced. Secondly, we briefly review existing surface parameterization techniques, especially for extensively studied parameterization algorithms for topological disk surfaces and parameterizations over regular domains for closed surfaces. Then we propose a polycube parameterization algorithm for closed surfaces with general topology. We develop an efficient optimization framework to minimize the angle and area distortion of the mapping. Its applications on surface meshing, inter-shape morphing and volumetric polycube mapping are also discussed
Repurpose 2D Animations for a VR Environment using BDH Shape Interpolation
Virtual Reality technology has spread rapidly in recent years. However, its growth risks ending soon due to the absence of quality content, except for few exceptions. We present an original framework that allows artists to use 2D characters and animations in a 3D Virtual Reality environment, in order to give an easier access to the production of content for the platform. In traditional platforms, 2D animation represents a more economic and immediate alternative to 3D. The challenge in adapting 2D characters to a 3D environment is to interpret the missing
depth information. A 2D character is actually flat, so there is not any depth information, and every body part is at the same level of the others. We exploit mesh interpolation, billboarding and parallax scrolling to simulate the depth between each body segment of the character. We have developed a prototype of the system, and extensive tests with a 2D animation production show the effectiveness of our framework
Repurpose 2D Character Animations for a VR Environment Using BDH Shape Interpolation.
Virtual Reality technology has spread rapidly in recent years.
However, its growth risks ending soon due to the absence of quality content, except for few exceptions. We present an original framework that allows artists to use 2D characters and animations in a 3D Virtual Reality environment, in order to give an easier access to the production of content for the platform. In traditional platforms, 2D animation represents a more economic and immediate alternative to 3D. The challenge in adapting 2D characters to a 3D environment is to interpret the missing depth information. A 2D character is actually flat, so there is not any depth information, and every body part is at the same level of the others. We exploit mesh interpolation, billboarding and parallax scrolling to simulate the depth between each body segment of the character. We have developed a prototype of the system, and extensive tests with a 2D animation production show the effectiveness of our framework
3D mesh metamorphosis from spherical parameterization for conceptual design
Engineering product design is an information intensive decision-making
process that consists of several phases including design specification
definition, design concepts generation, detailed design and analysis,
and manufacturing. Usually, generating geometry models for
visualization is a big challenge for early stage conceptual design.
Complexity of existing computer aided design packages constrains
participation of people with various backgrounds in the design
process. In addition, many design processes do not take advantage of
the rich amount of legacy information available for new concepts
creation.
The research presented here explores the use of advanced graphical
techniques to quickly and efficiently merge legacy information with
new design concepts to rapidly create new conceptual product designs.
3D mesh metamorphosis framework 3DMeshMorpher was created to
construct new models by navigating in a shape-space of registered
design models. The framework is composed of: i) a fast spherical
parameterization method to map a geometric model (genus-0) onto a unit
sphere; ii) a geometric feature identification and picking technique
based on 3D skeleton extraction; and iii) a LOD controllable 3D
remeshing scheme with spherical mesh subdivision based on the
developedspherical parameterization. This efficient software framework
enables designers to create numerous geometric concepts in real time
with a simple graphical user interface.
The spherical parameterization method is focused on closed genus-zero
meshes. It is based upon barycentric coordinates with convex boundary.
Unlike most existing similar approaches which deal with each vertex in
the mesh equally, the method developed in this research focuses
primarily on resolving overlapping areas, which helps speed the
parameterization process. The algorithm starts by normalizing the
source mesh onto a unit sphere and followed by some initial relaxation
via Gauss-Seidel iterations. Due to its emphasis on solving only
challenging overlapping regions, this parameterization process is much
faster than existing spherical mapping methods.
To ensure the correspondence of features from different models, we
introduce a skeleton based feature identification and picking method
for features alignment. Unlike traditional methods that align single
point for each feature, this method can provide alignments for
complete feature areas. This could help users to create more
reasonable intermediate morphing results with preserved topological
features. This skeleton featuring framework could potentially be
extended to automatic features alignment for geometries with similar
topologies. The skeleton extracted could also be applied for other
applications such as skeleton-based animations.
The 3D remeshing algorithm with spherical mesh subdivision is
developed to generate a common connectivity for different mesh models.
This method is derived from the concept of spherical mesh subdivision.
The local recursive subdivision can be set to match the desired LOD
(level of details) for source spherical mesh. Such LOD is controllable
and this allows various outputs with different resolutions. Such
recursive subdivision then follows by a triangular correction process
which ensures valid triangulations for the remeshing. And the final
mesh merging and reconstruction process produces the remeshing model
with desired LOD specified from user. Usually the final merged model
contains all the geometric details from each model with reasonable
amount of vertices, unlike other existing methods that result in big
amount of vertices in the merged model. Such multi-resolution outputs
with controllable LOD could also be applied in various other computer
graphics applications such as computer games
DSM-Net: Disentangled Structured Mesh Net for Controllable Generation of Fine Geometry
3D shape generation is a fundamental operation in computer graphics. While
significant progress has been made, especially with recent deep generative
models, it remains a challenge to synthesize high-quality geometric shapes with
rich detail and complex structure, in a controllable manner. To tackle this, we
introduce DSM-Net, a deep neural network that learns a disentangled structured
mesh representation for 3D shapes, where two key aspects of shapes, geometry
and structure, are encoded in a synergistic manner to ensure plausibility of
the generated shapes, while also being disentangled as much as possible. This
supports a range of novel shape generation applications with intuitive control,
such as interpolation of structure (geometry) while keeping geometry
(structure) unchanged. To achieve this, we simultaneously learn structure and
geometry through variational autoencoders (VAEs) in a hierarchical manner for
both, with bijective mappings at each level. In this manner we effectively
encode geometry and structure in separate latent spaces, while ensuring their
compatibility: the structure is used to guide the geometry and vice versa. At
the leaf level, the part geometry is represented using a conditional part VAE,
to encode high-quality geometric details, guided by the structure context as
the condition. Our method not only supports controllable generation
applications, but also produces high-quality synthesized shapes, outperforming
state-of-the-art methods
Anatomy Transfer
Characters with precise internal anatomy are important in film and visual effects, as well as in medical applications. We propose the first semi-automatic method for creating anatomical structures, such as bones, muscles, viscera and fat tissues. This is done by transferring a reference anatomical model from an input template to an arbitrary target character, only defined by its boundary representation (skin). The fat distribution of the target character needs to be specified. We can either infer this information from MRI data, or allow the users to express their creative intent through a new editing tool. The rest of our method runs automatically: it first transfers the bones to the target character, while maintaining their structure as much as possible. The bone layer, along with the target skin eroded using the fat thickness information, are then used to define a volume where we map the internal anatomy of the source model using harmonic (Laplacian) deformation. This way, we are able to quickly generate anatomical models for a large range of target characters, while maintaining anatomical constraints
- …