8 research outputs found

    Hand Gesture Recognition as Password to Open the Door with Camera and Convexity Defect Method

    Full text link
    Computer Vision is one of reasearch that gets a lot of attention with many applications. One of the application is the hand gesture recognition system. By using EmguCV, will be obtained camera images from webcam camera. The Pictures will be disegmented by using skin detection method for decrease noises in order to obtain the information needed. The final project of this system is to implement the convexity defect method for extracting images and recognize patterns of hand gesture that represent the characters A, B, C, D, and E. The parameters used in pattern recognition of hand gesture is the number and length of the line connecting the hull and defects derived from the pattern of hand gesture

    Hand Gesture Recognition as Password to Open The Door With Camera and Convexity Defect Method

    Get PDF
    Computer Vision is one of reasearch that gets a lot of attention with many applications. One of the application is the hand gesture recognition system. By using EmguCV, will be obtained camera images from webcam camera. The Pictures will be disegmented by using  skin detection method for decrease noises in order to obtain the information needed. The final project of this system is to implement the convexity defect method for extracting images and recognize patterns of hand gesture that represent the characters A, B, C, D, and E. The parameters used in pattern recognition of hand gesture is the number and length of the line connecting the hull and defects derived from the pattern of hand gesture

    Multi-cue Mid-level Grouping

    Full text link
    Abstract. Region proposal methods provide richer object hypotheses than sliding windows with dramatically fewer proposals, yet they still number in the thousands. This large quantity of proposals typically re-sults from a diversification step that propagates bottom-up ambiguity in the form of proposals to the next processing stage. In this paper, we take a complementary approach in which mid-level knowledge is used to re-solve bottom-up ambiguity at an earlier stage to allow a further reduction in the number of proposals. We present a method for generating regions using the mid-level grouping cues of closure and symmetry. In doing so, we combine mid-level cues that are typically used only in isolation, and leverage them to produce fewer but higher quality proposals. We empha-size that our model is mid-level by learning it on a limited number of objects while applying it to different objects, thus demonstrating that it is transferable to other objects. In our quantitative evaluation, we 1) establish the usefulness of each grouping cue by demonstrating incre-mental improvement, and 2) demonstrate improvement on two leading region proposal methods with a limited budget of proposals.

    Object Detection by Contour Segment Networks

    Get PDF

    Superedge grouping for object localization by combining appearance and shape information

    Full text link

    Salient closed boundary extraction with ratio contour

    Full text link

    Developmentally deep perceptual system for a humanoid robot

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2003.Includes bibliographical references (p. 139-152).This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.This thesis presents a perceptual system for a humanoid robot that integrates abilities such as object localization and recognition with the deeper developmental machinery required to forge those competences out of raw physical experiences. It shows that a robotic platform can build up and maintain a system for object localization, segmentation, and recognition, starting from very little. What the robot starts with is a direct solution to achieving figure/ground separation: it simply 'pokes around' in a region of visual ambiguity and watches what happens. If the arm passes through an area, that area is recognized as free space. If the arm collides with an object, causing it to move, the robot can use that motion to segment the object from the background. Once the robot can acquire reliable segmented views of objects, it learns from them, and from then on recognizes and segments those objects without further contact. Both low-level and high-level visual features can also be learned in this way, and examples are presented for both: orientation detection and affordance recognition, respectively. The motivation for this work is simple. Training on large corpora of annotated real-world data has proven crucial for creating robust solutions to perceptual problems such as speech recognition and face detection. But the powerful tools used during training of such systems are typically stripped away at deployment. Ideally they should remain, particularly for unstable tasks such as object detection, where the set of objects needed in a task tomorrow might be different from the set of objects needed today. The key limiting factor is access to training data, but as this thesis shows, that need not be a problem on a robotic platform that can actively probe its environment, and carry out experiments to resolve ambiguity.(cont.) This work is an instance of a general approach to learning a new perceptual judgment: find special situations in which the perceptual judgment is easy and study these situations to find correlated features that can be observed more generally.by Paul Michael Fitzpatrick.Ph.D

    From First Contact to Close Encounters: A Developmentally Deep Perceptual System for a Humanoid Robot

    Get PDF
    This thesis presents a perceptual system for a humanoid robot that integrates abilities such as object localization and recognition with the deeper developmental machinery required to forge those competences out of raw physical experiences. It shows that a robotic platform can build up and maintain a system for object localization, segmentation, and recognition, starting from very little. What the robot starts with is a direct solution to achieving figure/ground separation: it simply 'pokes around' in a region of visual ambiguity and watches what happens. If the arm passes through an area, that area is recognized as free space. If the arm collides with an object, causing it to move, the robot can use that motion to segment the object from the background. Once the robot can acquire reliable segmented views of objects, it learns from them, and from then on recognizes and segments those objects without further contact. Both low-level and high-level visual features can also be learned in this way, and examples are presented for both: orientation detection and affordance recognition, respectively. The motivation for this work is simple. Training on large corpora of annotated real-world data has proven crucial for creating robust solutions to perceptual problems such as speech recognition and face detection. But the powerful tools used during training of such systems are typically stripped away at deployment. Ideally they should remain, particularly for unstable tasks such as object detection, where the set of objects needed in a task tomorrow might be different from the set of objects needed today. The key limiting factor is access to training data, but as this thesis shows, that need not be a problem on a robotic platform that can actively probe its environment, and carry out experiments to resolve ambiguity. This work is an instance of a general approach to learning a new perceptual judgment: find special situations in which the perceptual judgment is easy and study these situations to find correlated features that can be observed more generally
    corecore