6,320 research outputs found
Compression for Smooth Shape Analysis
Most 3D shape analysis methods use triangular meshes to discretize both the
shape and functions on it as piecewise linear functions. With this
representation, shape analysis requires fine meshes to represent smooth shapes
and geometric operators like normals, curvatures, or Laplace-Beltrami
eigenfunctions at large computational and memory costs.
We avoid this bottleneck with a compression technique that represents a
smooth shape as subdivision surfaces and exploits the subdivision scheme to
parametrize smooth functions on that shape with a few control parameters. This
compression does not affect the accuracy of the Laplace-Beltrami operator and
its eigenfunctions and allow us to compute shape descriptors and shape
matchings at an accuracy comparable to triangular meshes but a fraction of the
computational cost.
Our framework can also compress surfaces represented by point clouds to do
shape analysis of 3D scanning data
Subdivision surface fitting to a dense mesh using ridges and umbilics
Fitting a sparse surface to approximate vast dense data is of interest for many applications: reverse engineering, recognition and compression, etc. The present work provides an approach to fit a Loop subdivision surface to a dense triangular mesh of arbitrary topology, whilst preserving and aligning the original features. The natural ridge-joined connectivity of umbilics and ridge-crossings is used as the connectivity of the control mesh for subdivision, so that the edges follow salient features on the surface. Furthermore, the chosen features and connectivity characterise the overall shape of the original mesh, since ridges capture extreme principal curvatures and ridges start and end at umbilics. A metric of Hausdorff distance including curvature vectors is proposed and implemented in a distance transform algorithm to construct the connectivity. Ridge-colour matching is introduced as a criterion for edge flipping to improve feature alignment. Several examples are provided to demonstrate the feature-preserving capability of the proposed approach
Towards multiple 3D bone surface identification and reconstruction using few 2D X-ray images for intraoperative applications
This article discusses a possible method to use a small number, e.g. 5, of conventional 2D X-ray images to reconstruct multiple 3D bone surfaces intraoperatively. Each bone’s edge contours in X-ray images are automatically identified. Sparse 3D landmark points of each bone are automatically reconstructed by pairing the 2D X-ray images. The reconstructed landmark point distribution on a surface is approximately optimal covering main characteristics of the surface. A statistical shape model, dense point distribution model (DPDM), is then used to fit the reconstructed optimal landmarks vertices to reconstruct a full surface of each bone separately. The reconstructed surfaces can then be visualised and manipulated by surgeons or used by surgical robotic systems
Multilinear Wavelets: A Statistical Shape Space for Human Faces
We present a statistical model for D human faces in varying expression,
which decomposes the surface of the face using a wavelet transform, and learns
many localized, decorrelated multilinear models on the resulting coefficients.
Using this model we are able to reconstruct faces from noisy and occluded D
face scans, and facial motion sequences. Accurate reconstruction of face shape
is important for applications such as tele-presence and gaming. The localized
and multi-scale nature of our model allows for recovery of fine-scale detail
while retaining robustness to severe noise and occlusion, and is
computationally efficient and scalable. We validate these properties
experimentally on challenging data in the form of static scans and motion
sequences. We show that in comparison to a global multilinear model, our model
better preserves fine detail and is computationally faster, while in comparison
to a localized PCA model, our model better handles variation in expression, is
faster, and allows us to fix identity parameters for a given subject.Comment: 10 pages, 7 figures; accepted to ECCV 201
ADAM: a general method for using various data types in asteroid reconstruction
We introduce ADAM, the All-Data Asteroid Modelling algorithm. ADAM is simple
and universal since it handles all disk-resolved data types (adaptive optics or
other images, interferometry, and range-Doppler radar data) in a uniform manner
via the 2D Fourier transform, enabling fast convergence in model optimization.
The resolved data can be combined with disk-integrated data (photometry). In
the reconstruction process, the difference between each data type is only a few
code lines defining the particular generalized projection from 3D onto a 2D
image plane. Occultation timings can be included as sparse silhouettes, and
thermal infrared data are efficiently handled with an approximate algorithm
that is sufficient in practice due to the dominance of the high-contrast
(boundary) pixels over the low-contrast (interior) ones. This is of particular
importance to the raw ALMA data that can be directly handled by ADAM without
having to construct the standard image. We study the reliability of the
inversion by using the independent shape supports of function series and
control-point surfaces. When other data are lacking, one can carry out fast
nonconvex lightcurve-only inversion, but any shape models resulting from it
should only be taken as illustrative global-scale ones.Comment: 11 pages, submitted to A&
- …