13,406 research outputs found
Salient Local 3D Features for 3D Shape Retrieval
In this paper we describe a new formulation for the 3D salient local features
based on the voxel grid inspired by the Scale Invariant Feature Transform
(SIFT). We use it to identify the salient keypoints (invariant points) on a 3D
voxelized model and calculate invariant 3D local feature descriptors at these
keypoints. We then use the bag of words approach on the 3D local features to
represent the 3D models for shape retrieval. The advantages of the method are
that it can be applied to rigid as well as to articulated and deformable 3D
models. Finally, this approach is applied for 3D Shape Retrieval on the McGill
articulated shape benchmark and then the retrieval results are presented and
compared to other methods.Comment: Three-Dimensional Imaging, Interaction, and Measurement. Edited by
Beraldin, J. Angelo; Cheok, Geraldine S.; McCarthy, Michael B.;
Neuschaefer-Rube, Ulrich; Baskurt, Atilla M.; McDowall, Ian E.; Dolinsky,
Margaret. Proceedings of the SPIE, Volume 7864, pp. 78640S-78640S-8 (2011).
Conference Location: San Francisco Airport, California, USA ISBN:
9780819484017 Date: 10 March 201
3-D Hand Pose Estimation from Kinect's Point Cloud Using Appearance Matching
We present a novel appearance-based approach for pose estimation of a human
hand using the point clouds provided by the low-cost Microsoft Kinect sensor.
Both the free-hand case, in which the hand is isolated from the surrounding
environment, and the hand-object case, in which the different types of
interactions are classified, have been considered. The hand-object case is
clearly the most challenging task having to deal with multiple tracks. The
approach proposed here belongs to the class of partial pose estimation where
the estimated pose in a frame is used for the initialization of the next one.
The pose estimation is obtained by applying a modified version of the Iterative
Closest Point (ICP) algorithm to synthetic models to obtain the rigid
transformation that aligns each model with respect to the input data. The
proposed framework uses a "pure" point cloud as provided by the Kinect sensor
without any other information such as RGB values or normal vector components.
For this reason, the proposed method can also be applied to data obtained from
other types of depth sensor, or RGB-D camera
- …