403 research outputs found

    Robust Temporally Coherent Laplacian Protrusion Segmentation of 3D Articulated Bodies

    Get PDF
    In motion analysis and understanding it is important to be able to fit a suitable model or structure to the temporal series of observed data, in order to describe motion patterns in a compact way, and to discriminate between them. In an unsupervised context, i.e., no prior model of the moving object(s) is available, such a structure has to be learned from the data in a bottom-up fashion. In recent times, volumetric approaches in which the motion is captured from a number of cameras and a voxel-set representation of the body is built from the camera views, have gained ground due to attractive features such as inherent view-invariance and robustness to occlusions. Automatic, unsupervised segmentation of moving bodies along entire sequences, in a temporally-coherent and robust way, has the potential to provide a means of constructing a bottom-up model of the moving body, and track motion cues that may be later exploited for motion classification. Spectral methods such as locally linear embedding (LLE) can be useful in this context, as they preserve "protrusions", i.e., high-curvature regions of the 3D volume, of articulated shapes, while improving their separation in a lower dimensional space, making them in this way easier to cluster. In this paper we therefore propose a spectral approach to unsupervised and temporally-coherent body-protrusion segmentation along time sequences. Volumetric shapes are clustered in an embedding space, clusters are propagated in time to ensure coherence, and merged or split to accommodate changes in the body's topology. Experiments on both synthetic and real sequences of dense voxel-set data are shown. This supports the ability of the proposed method to cluster body-parts consistently over time in a totally unsupervised fashion, its robustness to sampling density and shape quality, and its potential for bottom-up model constructionComment: 31 pages, 26 figure

    Diffeomorphic image registration with applications to deformation modelling between multiple data sets

    Get PDF
    Over last years, the diffeomorphic image registration algorithms have been successfully introduced into the field of the medical image analysis. At the same time, the particular usability of these techniques, in majority derived from the solid mathematical background, has been only quantitatively explored for the limited applications such as longitudinal studies on treatment quality, or diseases progression. The thesis considers the deformable image registration algorithms, seeking out those that maintain the medical correctness of the estimated dense deformation fields in terms of the preservation of the object and its neighbourhood topology, offer the reasonable computational complexity to satisfy time restrictions coming from the potential applications, and are able to cope with low quality data typically encountered in Adaptive Radiotherapy (ART). The research has led to the main emphasis being laid on the diffeomorphic image registration to achieve one-to-one mapping between images. This involves introduction of the log-domain parameterisation of the deformation field by its approximation via a stationary velocity field. A quantitative and qualitative examination of existing and newly proposed algorithms for pairwise deformable image registration presented in this thesis, shows that the log-Euclidean parameterisation can be successfully utilised in the biomedical applications. Although algorithms utilising the log-domain parameterisation have theoretical justification for maintaining diffeomorphism, in general, the deformation fields produced by them have similar properties as these estimated by classical methods. Having this in mind, the best compromise in terms of the quality of the deformation fields has been found for the consistent image registration framework. The experimental results suggest also that the image registration with the symmetrical warping of the input images outperforms the classical approaches, and simultaneously can be easily introduced to most known algorithms. Furthermore, the log-domain implicit group-wise image registration is proposed. By linking the various sets of images related to the different subjects, the proposed image registration approach establishes a common subject space and between-subject correspondences therein. Although the correspondences between groups of images can be found by performing the classic image registration, the reference image selection (not required in the proposed implementation), may lead to a biased mean image being estimated and the corresponding common subject space not adequate to represent the general properties of the data sets. The approaches to diffeomorphic image registration have been also utilised as the principal elements for estimating the movements of the organs in the pelvic area based on the dense deformation field prediction system driven by the partial information coming from the specific type of the measurements parameterised using the implicit surface representation, and recognising facial expressions where the stationary velocity fields are used as the facial expression descriptors. Both applications have been extensively evaluated based on the real representative data sets of three-dimensional volumes and two-dimensional images, and the obtained results indicate the practical usability of the proposed techniques

    A vision-based approach for human hand tracking and gesture recognition.

    Get PDF
    Hand gesture interface has been becoming an active topic of human-computer interaction (HCI). The utilization of hand gestures in human-computer interface enables human operators to interact with computer environments in a natural and intuitive manner. In particular, bare hand interpretation technique frees users from cumbersome, but typically required devices in communication with computers, thus offering the ease and naturalness in HCI. Meanwhile, virtual assembly (VA) applies virtual reality (VR) techniques in mechanical assembly. It constructs computer tools to help product engineers planning, evaluating, optimizing, and verifying the assembly of mechanical systems without the need of physical objects. However, traditional devices such as keyboards and mice are no longer adequate due to their inefficiency in handling three-dimensional (3D) tasks. Special VR devices, such as data gloves, have been mandatory in VA. This thesis proposes a novel gesture-based interface for the application of VA. It develops a hybrid approach to incorporate an appearance-based hand localization technique with a skin tone filter in support of gesture recognition and hand tracking in the 3D space. With this interface, bare hands become a convenient substitution of special VR devices. Experiment results demonstrate the flexibility and robustness introduced by the proposed method to HCI.Dept. of Computer Science. Paper copy at Leddy Library: Theses & Major Papers - Basement, West Bldg. / Call Number: Thesis2004 .L8. Source: Masters Abstracts International, Volume: 43-03, page: 0883. Adviser: Xiaobu Yuan. Thesis (M.Sc.)--University of Windsor (Canada), 2004

    Gait Analysis and Recognition for Automated Visual Surveillance

    No full text
    Human motion analysis has received a great attention from researchers in the last decade due to its potential use in different applications such as automated visual surveillance. This field of research focuses on the perception and recognition of human activities, including people identification. We explore a new approach for walking pedestrian detection in an unconstrained outdoor environment. The proposed algorithm is based on gait motion as the rhythm of the footprint pattern of walking people is considered the stable and characteristic feature for the classification of moving objects. The novelty of our approach is motivated by the latest research for people identification using gait. The experimental results confirmed the robustness of our method to discriminate between single walking subject, groups of people and vehicles with a successful detection rate of 100%. Furthermore, the results revealed the potential of our method to extend visual surveillance systems to recognize walking people. Furthermore, we propose a new approach to extract human joints (vertex positions) using a model-based method. The spatial templates describing the human gait motion are produced via gait analysis performed on data collected from manual labeling. The Elliptic Fourier Descriptors are used to represent the motion models in a parametric form. The heel strike data is exploited to reduce the dimensionality of the parametric models. People walk normal to the viewing plane, as major gait information is available in a sagittal view. The ankle, knee and hip joints are successfully extracted with high accuracy for indoor and outdoor data. In this way, we have established a baseline analysis which can be deployed in recognition, marker-less analysis and other areas. The experimental results confirmed the robustness of the model-based approach to recognise walking subjects with a correct classification rate of 95% using purely the dynamic features derived from the joint motion. Therefore, this confirms the early psychological theories claiming that the discriminative features for motion perception and people recognition are embedded in gait kinematics. Furthermore, to quantify the intrusive nature of gait recognition we explore the effects of the different covariate factors on the performance of gait recognition. The covariate factors include footwear, clothing, carrying conditions and walking speed. As far as the author can determine, this is the first major study of its kind in this field to analyse the covariate factors using a model-based method

    GRASP News Volume 9, Number 1

    Get PDF
    A report of the General Robotics and Active Sensory Perception (GRASP) Laboratory

    Reconstruction and analysis of dynamic shapes

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2010.Cataloged from PDF version of thesis.Includes bibliographical references (p. 122-141).Motion capture has revolutionized entertainment and influenced fields as diverse as the arts, sports, and medicine. This is despite the limitation that it tracks only a small set of surface points. On the other hand, 3D scanning techniques digitize complete surfaces of static objects, but are not applicable to moving shapes. I present methods that overcome both limitations, and can obtain the moving geometry of dynamic shapes (such as people and clothes in motion) and analyze it in order to advance computer animation. Further understanding of dynamic shapes will enable various industries to enhance virtual characters, advance robot locomotion, improve sports performance, and aid in medical rehabilitation, thus directly affecting our daily lives. My methods efficiently recover much of the expressiveness of dynamic shapes from the silhouettes alone. Furthermore, the reconstruction quality is greatly improved by including surface orientations (normals). In order to make reconstruction more practical, I strive to capture dynamic shapes in their natural environment, which I do by using hybrid inertial and acoustic sensors. After capture, the reconstructed dynamic shapes are analyzed in order to enhance their utility. My algorithms then allow animators to generate novel motions, such as transferring facial performances from one actor onto another using multi-linear models. The presented research provides some of the first and most accurate reconstructions of complex moving surfaces, and is among the few approaches that establish a relationship between different dynamic shapes.by Daniel Vlasic.Ph.D

    \u3cem\u3eGRASP News\u3c/em\u3e: Volume 9, Number 1

    Get PDF
    The past year at the GRASP Lab has been an exciting and productive period. As always, innovation and technical advancement arising from past research has lead to unexpected questions and fertile areas for new research. New robots, new mobile platforms, new sensors and cameras, and new personnel have all contributed to the breathtaking pace of the change. Perhaps the most significant change is the trend towards multi-disciplinary projects, most notable the multi-agent project (see inside for details on this, and all the other new and on-going projects). This issue of GRASP News covers the developments for the year 1992 and the first quarter of 1993

    Surface reflectance estimation from spatio-temporal subband statistics of moving object videos

    Get PDF
    Ankara : The Department of Electrical and Electronics Engineering and the Graduate School of Engineering and Science of Bilkent University, 2012.Thesis (Master's) -- Bilkent University, 2012.Includes bibliographical refences.Image motion can convey a broad range of object properties including 3D structure (structure from motion, SfM), animacy (biological motion), and its material. Our understanding of how the visual system may estimate complex properties such as surface reflectance or object rigidity from image motion is still limited. In order to reveal the neural mechanisms underlying surface material understanding, a natural point to begin with is to study the output of filters that mimic response properties of low level visual neurons to different classes of moving textures, such as patches of shiny and matte surfaces. To this end we designed spatio-temporal bandpass filters whose frequency response is the second order derivative of the Gaussian function. Those filters are generated towards eight orientations in three scales in the frequency domain. We computed responses of these filters to dynamic specular and matte textures. Specifically, we assessed the statistics of the resultant filter output histograms and calculated the mean, standard deviation, skewness and kurtosis of those histograms. We found that there were substantial differences in standard deviation and skewness of specular and matte texture subband histograms. To formally test whether these simple measurements can in fact predict surface material from image motion we developed a computer-assisted classifier based on these statistics. The results of the classification showed that, 75% of all movies are classified correctly, where the correct classification rate of shiny object movies is around 77% and the correct classification rate of matte object movies is around 71%. Next, we synthesized dynamic textures which resembled the subband statistics of videos of moving shiny and matte objects. Interestingly the appearance of these synthesized textures were neither shiny nor matte. Taken together our results indicate that there are differences in the spatio-temporal subband statistics of image motion generated by rotating matte and specular objects. While these differences may be utilized by the human brain during the perceptual process, our results on the synthesized textures suggest that the statistics may not be sufficient to judge the material qualities of an object.Külçe, OnurM.S
    corecore