7 research outputs found

    GPUMLib: Deep Learning SOM Library for Surface Reconstruction

    Get PDF
    The evolution of 3D scanning devices and innovation in computer processing power and storage capacity has sparked the revolution of producing big point-cloud datasets. This phenomenon has becoming an integral part of the sophisticated building design process especially in the era of 4th Industrial Revolution. The big point-cloud datasets have caused complexity in handling surface reconstruction and visualization since existing algorithms are not so readily available. In this context, the surface reconstruction intelligent algorithms need to be revolutionized to deal with big point-cloud datasets in tandem with the advancement of hardware processing power and storage capacity. In this study, we propose GPUMLib – deep learning library for self-organizing map (SOM-DLLib) to solve problems involving big point-cloud datasets from 3D scanning devices. The SOM-DLLib consists of multiple layers for reducing and optimizing those big point cloud datasets. The findings show the final objects are successfully reconstructed with optimized neighborhood representation and the performance becomes better as the size of point clouds increases

    Combining visual features and Growing Neural Gas networks for robotic 3D SLAM

    Get PDF
    The use of 3D data in mobile robotics provides valuable information about the robot’s environment. Traditionally, stereo cameras have been used as a low-cost 3D sensor. However, the lack of precision and texture for some surfaces suggests that the use of other 3D sensors could be more suitable. In this work, we examine the use of two sensors: an infrared SR4000 and a Kinect camera. We use a combination of 3D data obtained by these cameras, along with features obtained from 2D images acquired from these cameras, using a Growing Neural Gas (GNG) network applied to the 3D data. The goal is to obtain a robust egomotion technique. The GNG network is used to reduce the camera error. To calculate the egomotion, we test two methods for 3D registration. One is based on an iterative closest points algorithm, and the other employs random sample consensus. Finally, a simultaneous localization and mapping method is applied to the complete sequence to reduce the global error. The error from each sensor and the mapping results from the proposed method are examined.This work has been supported by Grant DPI2009-07144 and DPI2013-40534-R from Ministerio de Ciencia e Innovacion of the Spanish Government, University of Alicante Projects GRE09-16 and GRE10-35, and Valencian Government Project GV/2011/034

    3D model reconstruction using neural gas accelerated on GPU

    Get PDF
    In this work, we propose the use of the neural gas (NG), a neural network that uses an unsupervised Competitive Hebbian Learning (CHL) rule, to develop a reverse engineering process. This is a simple and accurate method to reconstruct objects from point clouds obtained from multiple overlapping views using low-cost sensors. In contrast to other methods that may need several stages that include downsampling, noise filtering and many other tasks, the NG automatically obtains the 3D model of the scanned objects. To demonstrate the validity of our proposal we tested our method with several models and performed a study of the neural network parameterization computing the quality of representation and also comparing results with other neural methods like growing neural gas and Kohonen maps or classical methods like Voxel Grid. We also reconstructed models acquired by low cost sensors that can be used in virtual and augmented reality environments for redesign or manipulation purposes. Since the NG algorithm has a strong computational cost we propose its acceleration. We have redesigned and implemented the NG learning algorithm to fit it onto Graphics Processing Units using CUDA. A speed-up of 180× faster is obtained compared to the sequential CPU version.This work was partially funded by the Spanish Government DPI2013-40534-R grant

    GPUMLib: deep learning SOM library for surface reconstruction

    Get PDF
    The evolution of 3D scanning devices and innovation in computer processing power and storage capacity has sparked the revolution of producing big point-cloud datasets. This phenomenon has becoming an integral part of the sophisticated building design process especially in the era of 4th Industrial Revolution. The big point-cloud datasets have caused complexity in handling surface reconstruction and visualization since existing algorithms are not so readily available. In this context, the surface reconstruction intelligent algorithms need to be revolutionized to deal with big point-cloud datasets in tandem with the advancement of hardware processing power and storage capacity. In this study, we propose GPUMLib - deep learning library for self-organizing map (SOM-DLLib) to solve problems involving big point-cloud datasets from 3D scanning devices. The SOM-DLLib consists of multiple layers for reducing and optimizing those big point cloud datasets. The findings show the final objects are successfully reconstructed with optimized neighborhood representation and the performance becomes better as the size of point clouds increases

    3D Surface Reconstruction of Noisy Point Clouds Using Growing Neural Gas: 3D Object/Scene Reconstruction

    Get PDF
    With the advent of low-cost 3D sensors and 3D printers, scene and object 3D surface reconstruction has become an important research topic in the last years. In this work, we propose an automatic (unsupervised) method for 3D surface reconstruction from raw unorganized point clouds acquired using low-cost 3D sensors. We have modified the growing neural gas network, which is a suitable model because of its flexibility, rapid adaptation and excellent quality of representation, to perform 3D surface reconstruction of different real-world objects and scenes. Some improvements have been made on the original algorithm considering colour and surface normal information of input data during the learning stage and creating complete triangular meshes instead of basic wire-frame representations. The proposed method is able to successfully create 3D faces online, whereas existing 3D reconstruction methods based on self-organizing maps required post-processing steps to close gaps and holes produced during the 3D reconstruction process. A set of quantitative and qualitative experiments were carried out to validate the proposed method. The method has been implemented and tested on real data, and has been found to be effective at reconstructing noisy point clouds obtained using low-cost 3D sensors.This work was partially funded by the Spanish Government DPI2013-40534-R Grant

    Reconstructing anatomical structures with growing self-reconstruction maps

    No full text
    corecore