2,819 research outputs found

    Furniture models learned from the WWW: using web catalogs to locate and categorize unknown furniture pieces in 3D laser scans

    Get PDF
    In this article, we investigate how autonomous robots can exploit the high quality information already available from the WWW concerning 3-D models of office furniture. Apart from the hobbyist effort in Google 3-D Warehouse, many companies providing office furnishings already have the models for considerable portions of the objects found in our workplaces and homes. In particular, we present an approach that allows a robot to learn generic models of typical office furniture using examples found in the Web. These generic models are then used by the robot to locate and categorize unknown furniture in real indoor environments

    Computer Vision Algorithms For An Automated Harvester

    Get PDF
    Image classification and segmentation are the two main important parts in the 3D vision system of a harvesting robot. Regarding the first part, the vision system aids in the real time identification of contaminated areas of the farm based on the damage identified using the robot’s camera. To solve the problem of identification, a fast and non-destructive method, Support Vector Machine (SVM), is applied to improve the recognition accuracy and efficiency of the robot. Initially, a median filter is applied to remove the inherent noise in the colored image. SIFT features of the image are then extracted and computed forming a vector, which is then quantized into visual words. Finally, the histogram of the frequency of each element in the visual vocabulary is created and fed into an SVM classifier, which categorizes the mushrooms as either class one or class two. Our preliminary results for image classification were promising and the experiments carried out on the data set highlight fast computation time and a high rate of accuracy, reaching over 90% using this method, which can be employed in real life scenario. As pertains to image Segmentation on the other hand, the vision system aids in real time identification of mushrooms but a stiff challenge is encountered in robot vision as the irregularly spaced mushrooms of uneven sizes often occlude each other due to the nature of mushroom growth in the growing environment. We address the issue of mushroom segmentation by following a multi-step process; the images are first segmented in HSV color space to locate the area of interest and then both the image gradient information from the area of interest and Hough transform methods are used to locate the center position and perimeter of each individual mushroom in XY plane. Afterwards, the depth map information given by Microsoft Kinect is employed to estimate the Z- depth of each individual mushroom, which is then being used to measure the distance between the robot end effector and center coordinate of each individual mushroom. We tested this algorithm under various environmental conditions and our segmentation results indicate this method provides sufficient computational speed and accuracy

    Model-based recognition of curves and surfaces using tactile data

    Get PDF
    Model-based object recognition has mostly been studied over inputs including images and range data. Though such data are global, cameras and range sensors are subject to occlusions and clutters, which often make recognition difficult and computationally expensive. In contrast, touch by a robot hand is free of occlusion and clutter issues, and recognition over tactile data can be more efficient.;In this thesis, we investigate model-based recognition of two and three dimensional curved objects from tactile data. The recognition of 2D objects is an invariant-based approach. We have derived differential and semi-differential invariants for quadratic curves and special cubic curves that are found in applications. These invariants, independent of translation and rotation, can be computed from local geometry of a curve. Invariants for quadratic curves are the functions in terms of the curvature and its derivative with respect to arc length. For cubic curves, the derived invariants also involve a slope in their expressions. Recognition of a curve reduces to invariant verification with its canonical parametric form determined along the way. In addition, the contact locations with the robot hand are found on the curve, thereby localizing it relative to the touch sensor. We have verified the correctness of all invariants by simulations. We have also shown that the shape parameters of the recognized curve can be recovered with small errors. The byproduct is a procedure that reliably estimates curvature and its derivative from real tactile data. The presented work distinguishes itself from traditional model-based recognition in its ability to simultaneously recognize and localize a shape from one of several classes, each consisting of a continuum of shapes, by the use of local data.;The recognition of 3D objects is based on registration and consists of two steps. First, a robotic hand with touch sensors samples data points on the object\u27s surface along three concurrent curves. The two principal curvatures at the curve intersection point are estimated and then used in a table lookup to find surface points that have similar local geometries. Next, starting at each such point, a local search is conducted to superpose the tactile data onto the surface model. Recognition of the model is based on the quality of this registration. The presented method can recognize algebraic as well as free-form surfaces, as demonstrated via simulations and robot experiments. One difference in the recognition of these two sets of shapes lies in the principal curvature estimation, which are calculated from the close forms and estimated through fitting, respectively. The other difference lies in data registration, which is carried out by nonlinear optimization and a greedy algorithm, respectively

    AirCode: Unobtrusive Physical Tags for Digital Fabrication

    Full text link
    We present AirCode, a technique that allows the user to tag physically fabricated objects with given information. An AirCode tag consists of a group of carefully designed air pockets placed beneath the object surface. These air pockets are easily produced during the fabrication process of the object, without any additional material or postprocessing. Meanwhile, the air pockets affect only the scattering light transport under the surface, and thus are hard to notice to our naked eyes. But, by using a computational imaging method, the tags become detectable. We present a tool that automates the design of air pockets for the user to encode information. AirCode system also allows the user to retrieve the information from captured images via a robust decoding algorithm. We demonstrate our tagging technique with applications for metadata embedding, robotic grasping, as well as conveying object affordances.Comment: ACM UIST 2017 Technical Paper

    Probabilistic Search for Object Segmentation and Recognition

    Get PDF
    The problem of searching for a model-based scene interpretation is analyzed within a probabilistic framework. Object models are formulated as generative models for range data of the scene. A new statistical criterion, the truncated object probability, is introduced to infer an optimal sequence of object hypotheses to be evaluated for their match to the data. The truncated probability is partly determined by prior knowledge of the objects and partly learned from data. Some experiments on sequence quality and object segmentation and recognition from stereo data are presented. The article recovers classic concepts from object recognition (grouping, geometric hashing, alignment) from the probabilistic perspective and adds insight into the optimal ordering of object hypotheses for evaluation. Moreover, it introduces point-relation densities, a key component of the truncated probability, as statistical models of local surface shape

    Probabilistic Search for Object Segmentation and Recognition

    Full text link
    The problem of searching for a model-based scene interpretation is analyzed within a probabilistic framework. Object models are formulated as generative models for range data of the scene. A new statistical criterion, the truncated object probability, is introduced to infer an optimal sequence of object hypotheses to be evaluated for their match to the data. The truncated probability is partly determined by prior knowledge of the objects and partly learned from data. Some experiments on sequence quality and object segmentation and recognition from stereo data are presented. The article recovers classic concepts from object recognition (grouping, geometric hashing, alignment) from the probabilistic perspective and adds insight into the optimal ordering of object hypotheses for evaluation. Moreover, it introduces point-relation densities, a key component of the truncated probability, as statistical models of local surface shape.Comment: 18 pages, 5 figure

    3D-POLY: A Robot Vision System for Recognizing Objects in Occluded Environments

    Get PDF
    The two factors that determine the time complexity associated with model-driven interpretation of range maps are: I) the particular strategy used for the generation of object hypotheses; and 2) the manner in which both the model and the sensed data are organized, data organization being a primary determinant of the efficiency of verification of a given hypothesis. In this report, we present 3D-POLY, a working system for recognizing objects in the presence of occlusion and against cluttered backgrounds. The time complexity of this system is only O(n2) for single object recognition, where n is the number of features on the object. The most novel aspect of this system is the manner in which the feature data are organized for the models. We use a data structure called the feature sphere for the purpose. We will present efficient algorithms for assigning a feature to its proper place on a feature sphere, and for extracting the neighbors of a given feature from the feature sphere representation. For hypothesis generation, we use local feature sets, a notion similar to those used before us by Bolles, Shirai and others. The combination of the feature sphere idea for streamlining verification and the local feature sets for hypothesis generation results in a system whose time complexity has a polynomial bound. In addition to recognizing objects in occluded environments, 3D-POLY also possesses model learning capability. Model learning consists of looking at a model object from different views and integrating the resulting information. The 3D-POLY system also contains utilities for range image segmentation and classification of scene surfaces

    Three Dimensional Measurement Using Fisheye Stereo Vision

    Get PDF
    corecore