7 research outputs found

    Face and Gesture Recognition for Human-Robot Interaction

    Get PDF

    A vision-based approach for human hand tracking and gesture recognition.

    Get PDF
    Hand gesture interface has been becoming an active topic of human-computer interaction (HCI). The utilization of hand gestures in human-computer interface enables human operators to interact with computer environments in a natural and intuitive manner. In particular, bare hand interpretation technique frees users from cumbersome, but typically required devices in communication with computers, thus offering the ease and naturalness in HCI. Meanwhile, virtual assembly (VA) applies virtual reality (VR) techniques in mechanical assembly. It constructs computer tools to help product engineers planning, evaluating, optimizing, and verifying the assembly of mechanical systems without the need of physical objects. However, traditional devices such as keyboards and mice are no longer adequate due to their inefficiency in handling three-dimensional (3D) tasks. Special VR devices, such as data gloves, have been mandatory in VA. This thesis proposes a novel gesture-based interface for the application of VA. It develops a hybrid approach to incorporate an appearance-based hand localization technique with a skin tone filter in support of gesture recognition and hand tracking in the 3D space. With this interface, bare hands become a convenient substitution of special VR devices. Experiment results demonstrate the flexibility and robustness introduced by the proposed method to HCI.Dept. of Computer Science. Paper copy at Leddy Library: Theses & Major Papers - Basement, West Bldg. / Call Number: Thesis2004 .L8. Source: Masters Abstracts International, Volume: 43-03, page: 0883. Adviser: Xiaobu Yuan. Thesis (M.Sc.)--University of Windsor (Canada), 2004

    Multiple cue integration for robust tracking in dynamic environments: application to video relighting

    Get PDF
    L'an脿lisi de moviment i seguiment d'objectes ha estat un dels pricipals focus d'atenci贸 en la comunitat de visi贸 per computador durant les dues darreres d猫cades. L'inter猫s per aquesta 脿rea de recerca resideix en el seu ample ventall d'aplicabilitat, que s'ext茅n des de tasques de navegaci贸 de vehicles aut貌noms i robots, fins a aplications en la ind煤stria de l'entreteniment i realitat virtual.Tot i que s'han aconseguit resultats espectaculars en problemes espec铆fics, el seguiment d'objectes continua essent un problema obert, ja que els m猫todes disponibles s贸n propensos a ser sensibles a diversos factors i condicions no estacion脿ries de l'entorn, com ara moviments impredictibles de l'objecte a seguir, canvis suaus o abruptes de la il路luminaci贸, proximitat d'objectes similars o fons confusos. Enfront aquests factors de confusi贸 la integraci贸 de m煤ltiples caracter铆stiques ha demostrat que permet millorar la robustesa dels algoritmes de seguiment. En els darrers anys, degut a la creixent capacitat de c脿lcul dels ordinadors, hi ha hagut un significatiu increment en el disseny de complexes sistemes de seguiment que consideren simult脿niament m煤ltiples caracter铆stiques de l'objecte. No obstant, la majoria d'aquests algoritmes estan basats enheur铆stiques i regles ad-hoc formulades per aplications espec铆fiques, fent-ne impossible l'extrapolaci贸 a noves condicions de l'entorn.En aquesta tesi proposem un marc probabil铆stic general per integrar el nombre de caracter铆stiques de l'objecte que siguin necess脿ries, permetent que interactuin m煤tuament per tal d'estimar-ne el seu estat amb precisi贸, i per tant, estimar amb precisi贸 la posici贸 de l'objecte que s'est脿 seguint. Aquest marc, s'utilitza posteriorment per dissenyar un algoritme de seguiment, que es valida en diverses seq眉猫ncies de v铆deo que contenen canvis abruptes de posici贸 i il路luminaci贸, camuflament de l'objecte i deformacions no r铆gides. Entre les caracter铆stiques que s'han utilitzat per representar l'objecte, cal destacar la paramatritzaci贸 robusta del color en un espai de color dependent de l'objecte, que permet distingir-lo del fons m茅s clarament que altres espais de color t铆picament ulitzats al llarg de la literatura.En la darrera part de la tesi dissenyem una t猫cnica per re-il路luminar tant escenes est脿tiques com en moviment, de les que s'en desconeix la geometria. La re-il路luminaci贸 es realitza amb un m猫tode 'basat en imatges', on la generaci贸 de les images de l'escena sota noves condicions d'il路luminaci贸 s'aconsegueix a partir de combinacions lineals d'un conjunt d'imatges de refer猫ncia pre-capturades, i que han estat generades il路luminant l'escena amb patrons de llum coneguts. Com que la posici贸 i intensitat de les fonts d'il.luminaci贸 que formen aquests patrons de llum es pot controlar, 茅s natural preguntar-nos: quina 茅s la manera m茅s 貌ptima d'il路luminar una escena per tal de reduir el nombre d'imatges de refer猫ncia? Demostrem que la millor manera d'il路luminar l'escena (茅s a dir, la que minimitza el nombre d'imatges de refer猫ncia) no 茅s utilitzant una seq眉猫ncia de fonts d'il路luminaci贸 puntuals, com es fa generalment, sin贸 a trav茅s d'una seq眉猫ncia de patrons de llum d'una base d'il路luminaci贸 depenent de l'objecte. 脡s important destacar que quan es re-il路luminen seq眉猫ncies de v铆deo, les imatges successives s'han d'alinear respecte a un sistema de coordenades com煤. Com que cada imatge ha estat generada per un patr贸 de llum diferent il路uminant l'escena, es produiran canvis d'il路luminaci贸 bruscos entre imatges de refer猫ncia consecutives. Sota aquestes circumst脿ncies, el m猫tode de seguiment proposat en aquesta tesi juga un paper fonamental. Finalment, presentem diversos resultats on re-il路luminem seq眉猫ncies de v铆deo reals d'objectes i cares d'actors en moviment. En cada cas, tot i que s'adquireix un 煤nic v铆deo, som capa莽os de re-il路luminar una i altra vegada, controlant la direcci贸 de la llum, la seva intensitat, i el color.Motion analysis and object tracking has been one of the principal focus of attention over the past two decades within the computer vision community. The interest of this research area lies in its wide range of applicability, extending from autonomous vehicle and robot navigation tasks, to entertainment and virtual reality applications.Even though impressive results have been obtained in specific problems, object tracking is still an open problem, since available methods are prone to be sensitive to several artifacts and non-stationary environment conditions, such as unpredictable target movements, gradual or abrupt changes of illumination, proximity of similar objects or cluttered backgrounds. Multiple cue integration has been proved to enhance the robustness of the tracking algorithms in front of such disturbances. In recent years, due to the increasing power of the computers, there has been a significant interest in building complex tracking systems which simultaneously consider multiple cues. However, most of these algorithms are based on heuristics and ad-hoc rules formulated for specific applications, making impossible to extrapolate them to new environment conditions.In this dissertation we propose a general probabilistic framework to integrate as many object features as necessary, permitting them to mutually interact in order to obtain a precise estimation of its state, and thus, a precise estimate of the target position. This framework is utilized to design a tracking algorithm, which is validated on several video sequences involving abrupt position and illumination changes, target camouflaging and non-rigid deformations. Among the utilized features to represent the target, it is important to point out the use of a robust parameterization of the target color in an object dependent colorspace which allows to distinguish the object from the background more clearly than other colorspaces commonly used in the literature.In the last part of the dissertation, we design an approach for relighting static and moving scenes with unknown geometry. The relighting is performed through an -image-based' methodology, where the rendering under new lighting conditions is achieved by linear combinations of a set of pre-acquired reference images of the scene illuminated by known light patterns. Since the placement and brightness of the light sources composing such light patterns can be controlled, it is natural to ask: what is the optimal way to illuminate the scene to reduce the number of reference images that are needed? We show that the best way to light the scene (i.e., the way that minimizes the number of reference images) is not using a sequence of single, compact light sources as is most commonly done, but rather to use a sequence of lighting patterns as given by an object-dependent lighting basis. It is important to note that when relighting video sequences, consecutive images need to be aligned with respect to a common coordinate frame. However, since each frame is generated by a different light pattern illuminating the scene, abrupt illumination changes between consecutive reference images are produced. Under these circumstances, the tracking framework designed in this dissertation plays a central role. Finally, we present several relighting results on real video sequences of moving objects, moving faces, and scenes containing both. In each case, although a single video clip was captured, we are able to relight again and again, controlling the lighting direction, extent, and color.Postprint (published version

    The Future of Humanoid Robots

    Get PDF
    This book provides state of the art scientific and engineering research findings and developments in the field of humanoid robotics and its applications. It is expected that humanoids will change the way we interact with machines, and will have the ability to blend perfectly into an environment already designed for humans. The book contains chapters that aim to discover the future abilities of humanoid robots by presenting a variety of integrated research in various scientific and engineering fields, such as locomotion, perception, adaptive behavior, human-robot interaction, neuroscience and machine learning. The book is designed to be accessible and practical, with an emphasis on useful information to those working in the fields of robotics, cognitive science, artificial intelligence, computational methods and other fields of science directly or indirectly related to the development and usage of future humanoid robots. The editor of the book has extensive R&D experience, patents, and publications in the area of humanoid robotics, and his experience is reflected in editing the content of the book
    corecore