7 research outputs found

    Object recognition in noisy RGB-D data using GNG

    Get PDF
    Object recognition in 3D scenes is a research field in which there is intense activity guided by the problems related to the use of 3D point clouds. Some of these problems are influenced by the presence of noise in the cloud that reduces the effectiveness of a recognition process. This work proposes a method for dealing with the noise present in point clouds by applying the growing neural gas (GNG) network filtering algorithm. This method is able to represent the input data with the desired number of neurons while preserving the topology of the input space. The GNG obtained results which were compared with a Voxel grid filter to determine the efficacy of our approach. Moreover, since a stage of the recognition process includes the detection of keypoints in a cloud, we evaluated different keypoint detectors to determine which one produces the best results in the selected pipeline. Experiments show how the GNG method yields better recognition results than other filtering algorithms when noise is present

    Stereo vision-based road estimation assisted by efficient planar patch calculation

    Get PDF

    Event-Driven Technologies for Reactive Motion Planning: Neuromorphic Stereo Vision and Robot Path Planning and Their Application on Parallel Hardware

    Get PDF
    Die Robotik wird immer mehr zu einem Schlüsselfaktor des technischen Aufschwungs. Trotz beeindruckender Fortschritte in den letzten Jahrzehnten, übertreffen Gehirne von Säugetieren in den Bereichen Sehen und Bewegungsplanung noch immer selbst die leistungsfähigsten Maschinen. Industrieroboter sind sehr schnell und präzise, aber ihre Planungsalgorithmen sind in hochdynamischen Umgebungen, wie sie für die Mensch-Roboter-Kollaboration (MRK) erforderlich sind, nicht leistungsfähig genug. Ohne schnelle und adaptive Bewegungsplanung kann sichere MRK nicht garantiert werden. Neuromorphe Technologien, einschließlich visueller Sensoren und Hardware-Chips, arbeiten asynchron und verarbeiten so raum-zeitliche Informationen sehr effizient. Insbesondere ereignisbasierte visuelle Sensoren sind konventionellen, synchronen Kameras bei vielen Anwendungen bereits überlegen. Daher haben ereignisbasierte Methoden ein großes Potenzial, schnellere und energieeffizientere Algorithmen zur Bewegungssteuerung in der MRK zu ermöglichen. In dieser Arbeit wird ein Ansatz zur flexiblen reaktiven Bewegungssteuerung eines Roboterarms vorgestellt. Dabei wird die Exterozeption durch ereignisbasiertes Stereosehen erreicht und die Pfadplanung ist in einer neuronalen Repräsentation des Konfigurationsraums implementiert. Die Multiview-3D-Rekonstruktion wird durch eine qualitative Analyse in Simulation evaluiert und auf ein Stereo-System ereignisbasierter Kameras übertragen. Zur Evaluierung der reaktiven kollisionsfreien Online-Planung wird ein Demonstrator mit einem industriellen Roboter genutzt. Dieser wird auch für eine vergleichende Studie zu sample-basierten Planern verwendet. Ergänzt wird dies durch einen Benchmark von parallelen Hardwarelösungen wozu als Testszenario Bahnplanung in der Robotik gewählt wurde. Die Ergebnisse zeigen, dass die vorgeschlagenen neuronalen Lösungen einen effektiven Weg zur Realisierung einer Robotersteuerung für dynamische Szenarien darstellen. Diese Arbeit schafft eine Grundlage für neuronale Lösungen bei adaptiven Fertigungsprozesse, auch in Zusammenarbeit mit dem Menschen, ohne Einbußen bei Geschwindigkeit und Sicherheit. Damit ebnet sie den Weg für die Integration von dem Gehirn nachempfundener Hardware und Algorithmen in die Industrierobotik und MRK

    Computational intelligence approaches to robotics, automation, and control [Volume guest editors]

    Get PDF
    No abstract available

    Using GNG to improve 3D feature extraction—Application to 6DoF egomotion

    Get PDF
    Several recent works deal with 3D data in mobile robotic problems, e.g. mapping or egomotion. Data comes from any kind of sensor such as stereo vision systems, time of flight cameras or 3D lasers, providing a huge amount of unorganized 3D data. In this paper, we describe an efficient method to build complete 3D models from a Growing Neural Gas (GNG). The GNG is applied to the 3D raw data and it reduces both the subjacent error and the number of points, keeping the topology of the 3D data. The GNG output is then used in a 3D feature extraction method. We have performed a deep study in which we quantitatively show that the use of GNG improves the 3D feature extraction method. We also show that our method can be applied to any kind of 3D data. The 3D features obtained are used as input in an Iterative Closest Point (ICP)-like method to compute the 6DoF movement performed by a mobile robot. A comparison with standard ICP is performed, showing that the use of GNG improves the results. Final results of 3D mapping from the egomotion calculated are also shown.This work has been partially supported by grant DPI2009-07144 from Ministerio de Ciencia e Innovacion of the Spanish Government and by the University of Alicante projects GRE09-16 and GRE10-35, and Valencia’s Government project GV/2011/034

    Three-dimensional image classification using hierarchical spatial decomposition: A study using retinal data

    Get PDF
    This thesis describes research conducted in the field of image mining especially volumetric image mining. The study investigates volumetric representation techniques based on hierarchical spatial decomposition to classify three-dimensional (3D) images. The aim of this study was to investigate the effectiveness of using hierarchical spatial decomposition coupled with regional homogeneity in the context of volumetric data representation. The proposed methods involve the following: (i) decomposition, (ii) representation, (iii) single feature vector generation and (iv) classifier generation. In the decomposition step, a given image (volume) is recursively decomposed until either homogeneous regions or a predefined maximum level are reached. For measuring the regional homogeneity, different critical functions are proposed. These critical functions are based on histograms of a given region. Once the image is decomposed, two representation methods are proposed: (i) to represent the decomposition using regions identified in the decomposition (region-based) or (ii) to represent the entire decomposition (whole image-based). The first method is based on individual regions, whereby each decomposed sub-volume (region) is represented in terms of different statistical and histogram-based techniques. Feature vector generation techniques are used to convert the set of feature vectors for each sub-volume into a single feature vector. In the whole image-based representation method, a tree is used to represent each image. Each node in the tree represents a region (sub-volume) using a single value and each edge describes the difference between the node and its parent node. A frequent sub-tree mining technique was adapted to identified a set of frequent sub-graphs. Selected sub-graphs are then used to build a feature vector for each image. In both cases, a standard classifier generator is applied, to the generated feature vectors, to model and predict the class of each image. Evaluation was conducted with respect to retinal optical coherence tomography images in terms of identifying Age-related Macular Degeneration (AMD). Two types of evaluation were used: (i) classification performance evaluation and (ii) statistical significance testing using ANalysis Of VAriance (ANOVA). The evaluation revealed that the proposed methods were effective for classifying 3D retinal images. It is consequently argued that the approaches are generic
    corecore